var/home/core/zuul-output/0000755000175000017500000000000015067442221014530 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067451021015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004464612515067451011017711 0ustar rootrootOct 02 09:27:12 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 09:27:12 crc restorecon[4670]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:12 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:13 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 09:27:14 crc restorecon[4670]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 09:27:15 crc kubenswrapper[5035]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:27:15 crc kubenswrapper[5035]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 09:27:15 crc kubenswrapper[5035]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:27:15 crc kubenswrapper[5035]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:27:15 crc kubenswrapper[5035]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 09:27:15 crc kubenswrapper[5035]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.716429 5035 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734180 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734233 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734240 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734246 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734251 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734257 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734264 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734269 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734275 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734280 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734285 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734291 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734297 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734302 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734307 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734313 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734321 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734329 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734337 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734344 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734351 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734356 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734361 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734366 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734378 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734383 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734388 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734393 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734398 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734403 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734408 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734414 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734420 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734428 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734436 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734444 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734450 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734457 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734463 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734469 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734475 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734483 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734489 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734495 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734501 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734509 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734515 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734521 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734527 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734562 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734572 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734580 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734590 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734597 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734602 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734607 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734611 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734616 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734621 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734626 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734631 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734636 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734640 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734646 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734652 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734658 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734663 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734669 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734674 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734679 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.734683 5035 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736626 5035 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736647 5035 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736657 5035 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736666 5035 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736675 5035 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736681 5035 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736690 5035 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736697 5035 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736740 5035 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736746 5035 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736752 5035 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736760 5035 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736766 5035 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736772 5035 flags.go:64] FLAG: --cgroup-root="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736778 5035 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736784 5035 flags.go:64] FLAG: --client-ca-file="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736789 5035 flags.go:64] FLAG: --cloud-config="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736795 5035 flags.go:64] FLAG: --cloud-provider="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736800 5035 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736808 5035 flags.go:64] FLAG: --cluster-domain="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736814 5035 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736820 5035 flags.go:64] FLAG: --config-dir="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736825 5035 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736831 5035 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736840 5035 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736847 5035 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736854 5035 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736861 5035 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736868 5035 flags.go:64] FLAG: --contention-profiling="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736875 5035 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736881 5035 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736887 5035 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736893 5035 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736905 5035 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736911 5035 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736916 5035 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736922 5035 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736929 5035 flags.go:64] FLAG: --enable-server="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736936 5035 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736944 5035 flags.go:64] FLAG: --event-burst="100" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736951 5035 flags.go:64] FLAG: --event-qps="50" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736957 5035 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736963 5035 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736968 5035 flags.go:64] FLAG: --eviction-hard="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736976 5035 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736981 5035 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736987 5035 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.736994 5035 flags.go:64] FLAG: --eviction-soft="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737000 5035 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737005 5035 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737012 5035 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737018 5035 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737024 5035 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737031 5035 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737037 5035 flags.go:64] FLAG: --feature-gates="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737051 5035 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737057 5035 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737063 5035 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737069 5035 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737075 5035 flags.go:64] FLAG: --healthz-port="10248" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737080 5035 flags.go:64] FLAG: --help="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737088 5035 flags.go:64] FLAG: --hostname-override="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737093 5035 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737099 5035 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737105 5035 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737110 5035 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737116 5035 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737122 5035 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737127 5035 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737133 5035 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737138 5035 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737144 5035 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737150 5035 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737155 5035 flags.go:64] FLAG: --kube-reserved="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737161 5035 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737167 5035 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737173 5035 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737178 5035 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737184 5035 flags.go:64] FLAG: --lock-file="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737190 5035 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737195 5035 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737201 5035 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737209 5035 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737217 5035 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737224 5035 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737230 5035 flags.go:64] FLAG: --logging-format="text" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737236 5035 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737242 5035 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737247 5035 flags.go:64] FLAG: --manifest-url="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737253 5035 flags.go:64] FLAG: --manifest-url-header="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737261 5035 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737267 5035 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737274 5035 flags.go:64] FLAG: --max-pods="110" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737280 5035 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737286 5035 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737291 5035 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737297 5035 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737303 5035 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737309 5035 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737314 5035 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737328 5035 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737334 5035 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737340 5035 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737346 5035 flags.go:64] FLAG: --pod-cidr="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737351 5035 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737360 5035 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737366 5035 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737372 5035 flags.go:64] FLAG: --pods-per-core="0" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737377 5035 flags.go:64] FLAG: --port="10250" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737383 5035 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737388 5035 flags.go:64] FLAG: --provider-id="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737394 5035 flags.go:64] FLAG: --qos-reserved="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737401 5035 flags.go:64] FLAG: --read-only-port="10255" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737408 5035 flags.go:64] FLAG: --register-node="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737415 5035 flags.go:64] FLAG: --register-schedulable="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737422 5035 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737435 5035 flags.go:64] FLAG: --registry-burst="10" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737442 5035 flags.go:64] FLAG: --registry-qps="5" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737449 5035 flags.go:64] FLAG: --reserved-cpus="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737458 5035 flags.go:64] FLAG: --reserved-memory="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737466 5035 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737472 5035 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737478 5035 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737483 5035 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737489 5035 flags.go:64] FLAG: --runonce="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737494 5035 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737501 5035 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737507 5035 flags.go:64] FLAG: --seccomp-default="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737513 5035 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737519 5035 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737525 5035 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737555 5035 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737561 5035 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737567 5035 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737573 5035 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737578 5035 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737585 5035 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737592 5035 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737603 5035 flags.go:64] FLAG: --system-cgroups="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737616 5035 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737631 5035 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737638 5035 flags.go:64] FLAG: --tls-cert-file="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737644 5035 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737653 5035 flags.go:64] FLAG: --tls-min-version="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737659 5035 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737665 5035 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737671 5035 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737678 5035 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737685 5035 flags.go:64] FLAG: --v="2" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737695 5035 flags.go:64] FLAG: --version="false" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737703 5035 flags.go:64] FLAG: --vmodule="" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737710 5035 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.737716 5035 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737849 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737856 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737862 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737868 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737873 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737879 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737885 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737890 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737895 5035 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737900 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737904 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737910 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737917 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737923 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737929 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737934 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737941 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737947 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737953 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737959 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737965 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737971 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737977 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737983 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737988 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737993 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.737998 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738003 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738008 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738015 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738021 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738028 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738033 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738039 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738044 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738049 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738055 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738066 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738072 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738077 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738082 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738087 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738091 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738096 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738101 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738106 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738112 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738116 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738121 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738126 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738130 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738135 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738140 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738145 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738150 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738154 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738159 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738165 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738172 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738177 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738182 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738189 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738194 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738201 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738206 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738212 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738217 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738222 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738229 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738237 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.738242 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.738260 5035 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.749275 5035 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.749326 5035 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749397 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749406 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749411 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749416 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749419 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749423 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749427 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749431 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749435 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749438 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749441 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749446 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749450 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749455 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749460 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749464 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749468 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749473 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749476 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749481 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749485 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749489 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749492 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749496 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749500 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749503 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749507 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749511 5035 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749516 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749520 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749525 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749545 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749550 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749556 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749565 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749571 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749576 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749581 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749586 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749590 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749594 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749599 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749603 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749607 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749612 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749616 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749620 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749624 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749630 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749637 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749642 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749647 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749653 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749657 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749662 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749666 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749671 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749675 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749678 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749683 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749687 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749691 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749695 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749699 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749703 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749707 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749711 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749715 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749718 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749722 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749726 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.749732 5035 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749852 5035 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749860 5035 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749865 5035 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749870 5035 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749874 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749878 5035 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749884 5035 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749891 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749895 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749903 5035 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749907 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749912 5035 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749916 5035 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749921 5035 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749925 5035 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749929 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749933 5035 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749937 5035 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749942 5035 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749946 5035 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749951 5035 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749955 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749959 5035 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749963 5035 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749968 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749971 5035 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749976 5035 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749980 5035 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749985 5035 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749989 5035 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749993 5035 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.749998 5035 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750002 5035 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750005 5035 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750010 5035 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750013 5035 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750017 5035 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750020 5035 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750024 5035 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750027 5035 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750031 5035 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750035 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750038 5035 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750042 5035 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750045 5035 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750049 5035 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750052 5035 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750056 5035 feature_gate.go:330] unrecognized feature gate: Example Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750060 5035 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750063 5035 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750067 5035 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750070 5035 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750074 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750077 5035 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750081 5035 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750084 5035 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750088 5035 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750091 5035 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750096 5035 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750100 5035 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750104 5035 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750108 5035 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750111 5035 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750115 5035 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750118 5035 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750122 5035 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750126 5035 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750130 5035 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750135 5035 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750139 5035 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 09:27:15 crc kubenswrapper[5035]: W1002 09:27:15.750143 5035 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.750150 5035 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.750309 5035 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.757449 5035 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.757650 5035 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.761217 5035 server.go:997] "Starting client certificate rotation" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.761253 5035 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.763145 5035 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-12 07:40:32.842059841 +0000 UTC Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.763225 5035 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2446h13m17.078838346s for next certificate rotation Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.810611 5035 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.814615 5035 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.846476 5035 log.go:25] "Validated CRI v1 runtime API" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.957070 5035 log.go:25] "Validated CRI v1 image API" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.959871 5035 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.985864 5035 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-09-22-30-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 09:27:15 crc kubenswrapper[5035]: I1002 09:27:15.985922 5035 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.015790 5035 manager.go:217] Machine: {Timestamp:2025-10-02 09:27:16.011343884 +0000 UTC m=+1.367687929 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a94415e2-ab3f-4649-975d-e580dfdd3e90 BootID:44cc1aef-5d31-4451-bb6a-5114e2dcb900 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b5:bc:0f Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b5:bc:0f Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:b6:00:6d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9a:92:b2 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ff:d4:62 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:52:51:20 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:36:b1:94:78:ee:be Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fe:d8:f3:52:2a:8d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.016087 5035 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.016318 5035 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.016678 5035 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.016878 5035 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.016913 5035 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.017100 5035 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.017109 5035 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.017658 5035 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.017686 5035 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.018276 5035 state_mem.go:36] "Initialized new in-memory state store" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.018386 5035 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.030443 5035 kubelet.go:418] "Attempting to sync node with API server" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.030511 5035 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.030566 5035 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.030582 5035 kubelet.go:324] "Adding apiserver pod source" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.030599 5035 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 09:27:16 crc kubenswrapper[5035]: W1002 09:27:16.056134 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:16 crc kubenswrapper[5035]: W1002 09:27:16.056135 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.056269 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.056302 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.057226 5035 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.058988 5035 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.083782 5035 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088146 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088187 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088196 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088205 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088218 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088227 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088236 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088249 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088260 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088269 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088282 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.088290 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.090988 5035 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.092003 5035 server.go:1280] "Started kubelet" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.092181 5035 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.092265 5035 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 09:27:16 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.094351 5035 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.096583 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.097718 5035 server.go:460] "Adding debug handlers to kubelet server" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.101660 5035 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.101721 5035 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.102113 5035 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.102139 5035 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.102234 5035 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 11:29:12.247235309 +0000 UTC Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.102315 5035 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2594h1m56.144925079s for next certificate rotation Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.102370 5035 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.102435 5035 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.103357 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="200ms" Oct 02 09:27:16 crc kubenswrapper[5035]: W1002 09:27:16.103429 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.103570 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.103861 5035 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.104020 5035 factory.go:55] Registering systemd factory Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.104171 5035 factory.go:221] Registration of the systemd container factory successfully Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.104989 5035 factory.go:153] Registering CRI-O factory Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.105038 5035 factory.go:221] Registration of the crio container factory successfully Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.105092 5035 factory.go:103] Registering Raw factory Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.105123 5035 manager.go:1196] Started watching for new ooms in manager Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.106219 5035 manager.go:319] Starting recovery of all containers Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127027 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127093 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127111 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127126 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127140 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127154 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127167 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127179 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127196 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127214 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127227 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127243 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127256 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127274 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127287 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127299 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127310 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127321 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127333 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127346 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127358 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127370 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127380 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127392 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127406 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127418 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127433 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127446 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127459 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127472 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127486 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127504 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127518 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127558 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127576 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127593 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127609 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127623 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127641 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127656 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127673 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127690 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127706 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127721 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127735 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127746 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127759 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127773 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127787 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127807 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127825 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.127839 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.128742 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.128831 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.128866 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.128994 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129026 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129283 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129363 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129385 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129417 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129431 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129451 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129466 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129480 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129500 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129514 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129547 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129562 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129576 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129599 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129612 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129634 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129647 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129659 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129680 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129692 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129711 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129725 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129738 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129761 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129774 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129797 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129812 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129826 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129846 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129860 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129890 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129908 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.129924 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132116 5035 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132148 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132169 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132195 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132208 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132228 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132241 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132253 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132272 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132285 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132298 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132320 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132332 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132352 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132366 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132405 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132421 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132446 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132465 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132489 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132507 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132526 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132556 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132575 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.129525 5035 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.154:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aa276ec109a95 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 09:27:16.091951765 +0000 UTC m=+1.448295830,LastTimestamp:2025-10-02 09:27:16.091951765 +0000 UTC m=+1.448295830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132592 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132606 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132625 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132639 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132651 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132669 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132691 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132740 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132753 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132768 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132786 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132799 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132815 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132833 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132846 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132864 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132876 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132889 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132906 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.132920 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.133124 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.134315 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.134346 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.134751 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.134783 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.134844 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135004 5035 manager.go:324] Recovery completed Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135408 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135475 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135492 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135510 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135527 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135560 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135584 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135598 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135614 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135629 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135643 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135659 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135674 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135689 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135714 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135731 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135745 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135759 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135774 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135788 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135802 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135815 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135828 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135842 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135857 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135871 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135885 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135901 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135916 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135931 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135947 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135960 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135973 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.135990 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136004 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136016 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136029 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136042 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136057 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136072 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136087 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136104 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136117 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136133 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136147 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136161 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136174 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136186 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136201 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136215 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136231 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136246 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136258 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136271 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136287 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136302 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136315 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136327 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136339 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136352 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136367 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136380 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136393 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136405 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136416 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136427 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136441 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136454 5035 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136468 5035 reconstruct.go:97] "Volume reconstruction finished" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.136479 5035 reconciler.go:26] "Reconciler: start to sync state" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.150492 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.152504 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.152584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.152601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.153552 5035 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.153569 5035 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.153590 5035 state_mem.go:36] "Initialized new in-memory state store" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.159547 5035 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.161623 5035 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.161675 5035 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.161711 5035 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.161765 5035 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 09:27:16 crc kubenswrapper[5035]: W1002 09:27:16.194064 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.194192 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.202755 5035 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.262757 5035 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.303050 5035 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.304929 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="400ms" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.309464 5035 policy_none.go:49] "None policy: Start" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.314157 5035 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.314213 5035 state_mem.go:35] "Initializing new in-memory state store" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.403188 5035 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.462901 5035 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.504184 5035 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.520881 5035 manager.go:334] "Starting Device Plugin manager" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.521280 5035 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.521322 5035 server.go:79] "Starting device plugin registration server" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.521963 5035 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.521987 5035 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.522265 5035 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.522391 5035 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.522405 5035 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.533008 5035 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.622347 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.623837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.623887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.623903 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.623940 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.624649 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.706483 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="800ms" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.825207 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.827110 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.827192 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.827216 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.827261 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:16 crc kubenswrapper[5035]: E1002 09:27:16.828098 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.864036 5035 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.864203 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.866083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.866161 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.866187 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.866434 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.866675 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.866712 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.867788 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.867817 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.867827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.867957 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.868000 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.868012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.868179 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.868410 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.868448 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.869376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.869410 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.869425 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.869998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.870025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.870035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.870135 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.870630 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.870684 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.871468 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.871498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.871515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.871653 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.872149 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.872193 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.872726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.872759 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.872774 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.873407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.873432 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:16 crc kubenswrapper[5035]: I1002 09:27:16.873445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.873609 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.873639 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.874059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.874082 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.874105 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.874465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.874482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.874494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:17 crc kubenswrapper[5035]: W1002 09:27:16.875954 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:17 crc kubenswrapper[5035]: E1002 09:27:16.876033 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946455 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946502 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946565 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946612 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946637 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946666 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946782 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946851 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946902 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946948 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.946990 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.947064 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.947146 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.947200 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:16.947245 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049017 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049054 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049074 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049094 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049114 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049131 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049148 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049164 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049184 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049202 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049220 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049237 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049254 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049271 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049287 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049685 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049733 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049755 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049776 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049820 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049842 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049863 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049882 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049899 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049897 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049945 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049934 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049918 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.049982 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.050005 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.086928 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.097339 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.099277 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.619130 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.619755 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: E1002 09:27:17.620041 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="1.6s" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.620764 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:17 crc kubenswrapper[5035]: I1002 09:27:17.620996 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:17 crc kubenswrapper[5035]: W1002 09:27:17.621923 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:17 crc kubenswrapper[5035]: E1002 09:27:17.622051 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.105827 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:18 crc kubenswrapper[5035]: E1002 09:27:18.106054 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.106435 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.106986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.107039 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.107064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.107113 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.107191 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:18 crc kubenswrapper[5035]: E1002 09:27:18.107470 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:18 crc kubenswrapper[5035]: E1002 09:27:18.107653 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.560738 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-dfdbd1cdf605d0d8ca204ea7c332a3279c80f2ecfedbf2994108ab1908eae3ab WatchSource:0}: Error finding container dfdbd1cdf605d0d8ca204ea7c332a3279c80f2ecfedbf2994108ab1908eae3ab: Status 404 returned error can't find the container with id dfdbd1cdf605d0d8ca204ea7c332a3279c80f2ecfedbf2994108ab1908eae3ab Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.561938 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5cb579d56790b86a637eaf7f54e1c8855a5f131e3871d50eb13e43a8320df16e WatchSource:0}: Error finding container 5cb579d56790b86a637eaf7f54e1c8855a5f131e3871d50eb13e43a8320df16e: Status 404 returned error can't find the container with id 5cb579d56790b86a637eaf7f54e1c8855a5f131e3871d50eb13e43a8320df16e Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.564330 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-0d89205bafab32ad7ee0e8785b5b2b803d304f85c2081676bc8622ff3cbda669 WatchSource:0}: Error finding container 0d89205bafab32ad7ee0e8785b5b2b803d304f85c2081676bc8622ff3cbda669: Status 404 returned error can't find the container with id 0d89205bafab32ad7ee0e8785b5b2b803d304f85c2081676bc8622ff3cbda669 Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.566830 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-6a5f14dfe09edd8541535a74ad8f3d300805e01ce27cefe8cc5fced0d9dc8b60 WatchSource:0}: Error finding container 6a5f14dfe09edd8541535a74ad8f3d300805e01ce27cefe8cc5fced0d9dc8b60: Status 404 returned error can't find the container with id 6a5f14dfe09edd8541535a74ad8f3d300805e01ce27cefe8cc5fced0d9dc8b60 Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.567450 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-50838ba88bde617d58be97657953d735c0a005882d67fc1c605ca1d3673defe8 WatchSource:0}: Error finding container 50838ba88bde617d58be97657953d735c0a005882d67fc1c605ca1d3673defe8: Status 404 returned error can't find the container with id 50838ba88bde617d58be97657953d735c0a005882d67fc1c605ca1d3673defe8 Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.625641 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6a5f14dfe09edd8541535a74ad8f3d300805e01ce27cefe8cc5fced0d9dc8b60"} Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.627098 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dfdbd1cdf605d0d8ca204ea7c332a3279c80f2ecfedbf2994108ab1908eae3ab"} Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.908070 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.910156 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.910237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.910256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.910303 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:18 crc kubenswrapper[5035]: E1002 09:27:18.911146 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.948285 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5cb579d56790b86a637eaf7f54e1c8855a5f131e3871d50eb13e43a8320df16e"} Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.950313 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"50838ba88bde617d58be97657953d735c0a005882d67fc1c605ca1d3673defe8"} Oct 02 09:27:18 crc kubenswrapper[5035]: I1002 09:27:18.951856 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0d89205bafab32ad7ee0e8785b5b2b803d304f85c2081676bc8622ff3cbda669"} Oct 02 09:27:18 crc kubenswrapper[5035]: W1002 09:27:18.995164 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:18 crc kubenswrapper[5035]: E1002 09:27:18.995309 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:19 crc kubenswrapper[5035]: I1002 09:27:19.098317 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:19 crc kubenswrapper[5035]: E1002 09:27:19.221228 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="3.2s" Oct 02 09:27:19 crc kubenswrapper[5035]: W1002 09:27:19.677068 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:19 crc kubenswrapper[5035]: E1002 09:27:19.677189 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:20 crc kubenswrapper[5035]: I1002 09:27:20.098333 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:20 crc kubenswrapper[5035]: W1002 09:27:20.435177 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:20 crc kubenswrapper[5035]: E1002 09:27:20.435281 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:20 crc kubenswrapper[5035]: I1002 09:27:20.511762 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:20 crc kubenswrapper[5035]: I1002 09:27:20.514936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:20 crc kubenswrapper[5035]: I1002 09:27:20.515011 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:20 crc kubenswrapper[5035]: I1002 09:27:20.515035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:20 crc kubenswrapper[5035]: I1002 09:27:20.515082 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:20 crc kubenswrapper[5035]: E1002 09:27:20.515901 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:21 crc kubenswrapper[5035]: W1002 09:27:21.093660 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:21 crc kubenswrapper[5035]: E1002 09:27:21.093763 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:21 crc kubenswrapper[5035]: I1002 09:27:21.097987 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:22 crc kubenswrapper[5035]: I1002 09:27:22.098022 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:22 crc kubenswrapper[5035]: E1002 09:27:22.311446 5035 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.154:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aa276ec109a95 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 09:27:16.091951765 +0000 UTC m=+1.448295830,LastTimestamp:2025-10-02 09:27:16.091951765 +0000 UTC m=+1.448295830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 09:27:22 crc kubenswrapper[5035]: E1002 09:27:22.422506 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="6.4s" Oct 02 09:27:22 crc kubenswrapper[5035]: W1002 09:27:22.976469 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:22 crc kubenswrapper[5035]: E1002 09:27:22.976519 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:22 crc kubenswrapper[5035]: W1002 09:27:22.979551 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:22 crc kubenswrapper[5035]: E1002 09:27:22.979600 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:23 crc kubenswrapper[5035]: I1002 09:27:23.097911 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:23 crc kubenswrapper[5035]: I1002 09:27:23.716684 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:23 crc kubenswrapper[5035]: I1002 09:27:23.718589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:23 crc kubenswrapper[5035]: I1002 09:27:23.718650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:23 crc kubenswrapper[5035]: I1002 09:27:23.718673 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:23 crc kubenswrapper[5035]: I1002 09:27:23.718712 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:23 crc kubenswrapper[5035]: E1002 09:27:23.719362 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:24 crc kubenswrapper[5035]: I1002 09:27:24.098226 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:24 crc kubenswrapper[5035]: W1002 09:27:24.590275 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:24 crc kubenswrapper[5035]: E1002 09:27:24.590356 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:24 crc kubenswrapper[5035]: W1002 09:27:24.770299 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:24 crc kubenswrapper[5035]: E1002 09:27:24.770415 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:25 crc kubenswrapper[5035]: I1002 09:27:25.098242 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:26 crc kubenswrapper[5035]: I1002 09:27:26.097925 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:26 crc kubenswrapper[5035]: E1002 09:27:26.533380 5035 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:27:27 crc kubenswrapper[5035]: I1002 09:27:27.098284 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:28 crc kubenswrapper[5035]: I1002 09:27:28.097444 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:28 crc kubenswrapper[5035]: E1002 09:27:28.823908 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="7s" Oct 02 09:27:29 crc kubenswrapper[5035]: I1002 09:27:29.098183 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:30 crc kubenswrapper[5035]: W1002 09:27:30.054802 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:30 crc kubenswrapper[5035]: E1002 09:27:30.054947 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.098141 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.119709 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.121345 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.121413 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.121429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.121467 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:30 crc kubenswrapper[5035]: E1002 09:27:30.122124 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:30 crc kubenswrapper[5035]: W1002 09:27:30.256700 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:30 crc kubenswrapper[5035]: E1002 09:27:30.256831 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.984007 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783"} Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.988933 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc"} Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.990831 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18"} Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.990971 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.992331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948"} Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.992609 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.992845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.992860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:30 crc kubenswrapper[5035]: I1002 09:27:30.994464 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927"} Oct 02 09:27:31 crc kubenswrapper[5035]: I1002 09:27:31.097418 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:31 crc kubenswrapper[5035]: I1002 09:27:31.999346 5035 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927" exitCode=0 Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:31.999499 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927"} Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:31.999627 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.001811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.001856 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.001868 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.002881 5035 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783" exitCode=0 Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.002952 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783"} Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.002996 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.004182 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.004234 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.004250 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.006627 5035 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc" exitCode=0 Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.006674 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc"} Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.006771 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.007861 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.007907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.007925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.009182 5035 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18" exitCode=0 Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.009255 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.009279 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18"} Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.010032 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.010596 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.010637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.010651 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.012094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.012132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.012148 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.012598 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722"} Oct 02 09:27:32 crc kubenswrapper[5035]: I1002 09:27:32.097397 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:32 crc kubenswrapper[5035]: E1002 09:27:32.312856 5035 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.154:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186aa276ec109a95 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 09:27:16.091951765 +0000 UTC m=+1.448295830,LastTimestamp:2025-10-02 09:27:16.091951765 +0000 UTC m=+1.448295830,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.018418 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d"} Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.018459 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.019739 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.019793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.019804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.023573 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9"} Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.025897 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63"} Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.028364 5035 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69" exitCode=0 Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.028455 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69"} Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.030818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3"} Oct 02 09:27:33 crc kubenswrapper[5035]: I1002 09:27:33.098604 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.037315 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5"} Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.042749 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f"} Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.043035 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5"} Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.046544 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658"} Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.046646 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.046728 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.046647 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.048264 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.048314 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.048315 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.048325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.048350 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.048503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.049118 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.049156 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.049171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:34 crc kubenswrapper[5035]: I1002 09:27:34.097704 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:34 crc kubenswrapper[5035]: W1002 09:27:34.286183 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:34 crc kubenswrapper[5035]: E1002 09:27:34.286264 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.051452 5035 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205" exitCode=0 Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.051515 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205"} Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.051758 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.053226 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.053258 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.053269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.055215 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681"} Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.055266 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.056498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.056582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.056595 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.060260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7de60c653a542841c4a9cbe1629cdb635dcc69a09338de12639ff664f9c40585"} Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.060300 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a"} Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.060271 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.060348 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.061513 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.061555 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.061565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.062150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.062184 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.062196 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.098016 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.262515 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.262977 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Oct 02 09:27:35 crc kubenswrapper[5035]: I1002 09:27:35.263059 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": dial tcp 192.168.126.11:6443: connect: connection refused" Oct 02 09:27:35 crc kubenswrapper[5035]: E1002 09:27:35.825241 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.154:6443: connect: connection refused" interval="7s" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.064871 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5811facbcc46f163c5dc71d7fc6095a509746b50764410ed99c94ebb24c317ba"} Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.064938 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.064966 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.064938 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.065059 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.066035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.066053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.066072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.066084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.066085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.066171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:36 crc kubenswrapper[5035]: I1002 09:27:36.097503 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:36 crc kubenswrapper[5035]: E1002 09:27:36.533624 5035 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.070691 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.070685 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd12dcbe766e84270752eb65d14f3e5881640d0e46d94c6c2d695edc83b6915e"} Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.070767 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.071726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.071752 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.071761 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.072128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.072191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.072209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.097353 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.122857 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.124928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.125001 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.125025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.125072 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:37 crc kubenswrapper[5035]: E1002 09:27:37.125870 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.154:6443: connect: connection refused" node="crc" Oct 02 09:27:37 crc kubenswrapper[5035]: W1002 09:27:37.172935 5035 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:37 crc kubenswrapper[5035]: E1002 09:27:37.173085 5035 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.154:6443: connect: connection refused" logger="UnhandledError" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.619683 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.619877 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.621916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.621995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.622013 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:37 crc kubenswrapper[5035]: I1002 09:27:37.989742 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:38 crc kubenswrapper[5035]: I1002 09:27:38.077458 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f483770dacbda745acb21802f775090ff51df5bc932518311b645c8539c2d5f9"} Oct 02 09:27:38 crc kubenswrapper[5035]: I1002 09:27:38.077495 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:38 crc kubenswrapper[5035]: I1002 09:27:38.078836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:38 crc kubenswrapper[5035]: I1002 09:27:38.078887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:38 crc kubenswrapper[5035]: I1002 09:27:38.078907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:38 crc kubenswrapper[5035]: I1002 09:27:38.097885 5035 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.154:6443: connect: connection refused Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.081454 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.088495 5035 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7de60c653a542841c4a9cbe1629cdb635dcc69a09338de12639ff664f9c40585" exitCode=255 Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.088586 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7de60c653a542841c4a9cbe1629cdb635dcc69a09338de12639ff664f9c40585"} Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.088764 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.089613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.089659 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.089675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.090399 5035 scope.go:117] "RemoveContainer" containerID="7de60c653a542841c4a9cbe1629cdb635dcc69a09338de12639ff664f9c40585" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.092819 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3dd669b07e45dd36d35fbe716218112dd79d06ef543b47b00108be49cfa39efa"} Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.354362 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.354584 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.357178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.357257 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.357284 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:39 crc kubenswrapper[5035]: I1002 09:27:39.363197 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.096746 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.099109 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b"} Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.099368 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.100870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.100905 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.100919 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.104190 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8a6e2066bf1530e2ef3f455e83a2822b8db1d1eaccdb3c199d4748f975771957"} Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.104249 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.104200 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.104388 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.105122 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.105163 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.105181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.105594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.105655 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.105680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.109003 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:40 crc kubenswrapper[5035]: I1002 09:27:40.598971 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.066182 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.106867 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.106867 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.107887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.107930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.107944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.108136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.108170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:41 crc kubenswrapper[5035]: I1002 09:27:41.108183 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.108861 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.108966 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.110090 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.110131 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.110144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.110158 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.110185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:42 crc kubenswrapper[5035]: I1002 09:27:42.110196 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:43 crc kubenswrapper[5035]: I1002 09:27:43.599576 5035 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 09:27:43 crc kubenswrapper[5035]: I1002 09:27:43.599682 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 09:27:43 crc kubenswrapper[5035]: I1002 09:27:43.774735 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 09:27:43 crc kubenswrapper[5035]: I1002 09:27:43.775035 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:43 crc kubenswrapper[5035]: I1002 09:27:43.776957 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:43 crc kubenswrapper[5035]: I1002 09:27:43.777031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:43 crc kubenswrapper[5035]: I1002 09:27:43.777058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:44 crc kubenswrapper[5035]: I1002 09:27:44.126468 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:44 crc kubenswrapper[5035]: I1002 09:27:44.128037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:44 crc kubenswrapper[5035]: I1002 09:27:44.128105 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:44 crc kubenswrapper[5035]: I1002 09:27:44.128129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:44 crc kubenswrapper[5035]: I1002 09:27:44.128172 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:27:44 crc kubenswrapper[5035]: I1002 09:27:44.999316 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:45 crc kubenswrapper[5035]: I1002 09:27:44.999614 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:45 crc kubenswrapper[5035]: I1002 09:27:45.001185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:45 crc kubenswrapper[5035]: I1002 09:27:45.001225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:45 crc kubenswrapper[5035]: I1002 09:27:45.001237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:46 crc kubenswrapper[5035]: E1002 09:27:46.534494 5035 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 09:27:48 crc kubenswrapper[5035]: I1002 09:27:48.573292 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 09:27:48 crc kubenswrapper[5035]: I1002 09:27:48.573766 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 09:27:48 crc kubenswrapper[5035]: I1002 09:27:48.579419 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 09:27:48 crc kubenswrapper[5035]: I1002 09:27:48.579500 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.271760 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.272004 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.272450 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.272514 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.273825 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.273893 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.273908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.276730 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:27:50 crc kubenswrapper[5035]: I1002 09:27:50.847992 5035 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.109945 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.110143 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.111458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.111572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.111594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.124427 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.132199 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.132198 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.132489 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.132524 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.133321 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.133376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.133325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.133413 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.133426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:51 crc kubenswrapper[5035]: I1002 09:27:51.133392 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:53 crc kubenswrapper[5035]: E1002 09:27:53.570286 5035 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="7s" Oct 02 09:27:53 crc kubenswrapper[5035]: E1002 09:27:53.573505 5035 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 09:27:53 crc kubenswrapper[5035]: I1002 09:27:53.575100 5035 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 09:27:53 crc kubenswrapper[5035]: I1002 09:27:53.575446 5035 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 09:27:53 crc kubenswrapper[5035]: I1002 09:27:53.600005 5035 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 09:27:53 crc kubenswrapper[5035]: I1002 09:27:53.600092 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 09:27:53 crc kubenswrapper[5035]: I1002 09:27:53.988471 5035 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48656->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 09:27:53 crc kubenswrapper[5035]: I1002 09:27:53.988565 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:48656->192.168.126.11:17697: read: connection reset by peer" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.142643 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.143067 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.144942 5035 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b" exitCode=255 Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.144962 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b"} Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.145018 5035 scope.go:117] "RemoveContainer" containerID="7de60c653a542841c4a9cbe1629cdb635dcc69a09338de12639ff664f9c40585" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.145154 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.146065 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.146592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.146611 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.147264 5035 scope.go:117] "RemoveContainer" containerID="1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b" Oct 02 09:27:54 crc kubenswrapper[5035]: E1002 09:27:54.147457 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:27:54 crc kubenswrapper[5035]: I1002 09:27:54.700338 5035 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.127737 5035 apiserver.go:52] "Watching apiserver" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.148901 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.178833 5035 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.179310 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-4hwst","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-cmx5n","openshift-multus/multus-8rpfb","openshift-machine-config-operator/machine-config-daemon-k6b5v","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-j6gm8"] Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.179711 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.179785 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.179856 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.180035 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.180157 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.180287 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.180408 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.180683 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.180718 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.180864 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.180886 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.181162 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.181235 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.181255 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.183890 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.184272 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.184301 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.189439 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.189445 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.189814 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.190522 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.190619 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.190644 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.190825 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.191145 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.191222 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.191798 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.191923 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200018 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200291 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200335 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200335 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200413 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200470 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200503 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200577 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.200883 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.201490 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.201766 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.201857 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.202094 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.202094 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.202212 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.202433 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.202492 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.203417 5035 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.215378 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.225835 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.241179 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.252985 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.262382 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.271368 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.280254 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284316 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284377 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284419 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284442 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284498 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284518 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284561 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284579 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284599 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284839 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.284922 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285151 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285260 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285457 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285581 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285607 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285646 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285680 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285700 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285699 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285746 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285764 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285783 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285800 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285820 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285872 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285894 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285913 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285942 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285966 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.285989 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286011 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286034 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286066 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286088 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286111 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286153 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286193 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286210 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286224 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286226 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286295 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286325 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286349 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286371 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286395 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286418 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286439 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286463 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286484 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286505 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286546 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286576 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286601 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286627 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286649 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286687 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286712 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286738 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286763 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286786 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286808 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286847 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286872 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286901 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286924 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286957 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286980 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287005 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287028 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287051 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287074 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287094 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287116 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287143 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287171 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287192 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287214 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287244 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287268 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287288 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287311 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287332 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287351 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287374 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287397 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287423 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287443 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287465 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287489 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287515 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287558 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287610 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287635 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287657 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287687 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287706 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287728 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287750 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287772 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287792 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287826 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287847 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287873 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.287980 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288006 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288027 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288049 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288070 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288092 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286363 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286395 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286485 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286510 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286592 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286658 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286668 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286733 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.286860 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288370 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288508 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288862 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.288878 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289081 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289292 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289509 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289549 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289632 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289664 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289687 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289732 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289750 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289769 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289769 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289793 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289817 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289837 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289856 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289840 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.289879 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290046 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290071 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290022 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290096 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290130 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290151 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290175 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290203 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290236 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290263 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290290 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290319 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290347 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290372 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290396 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290426 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290456 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290482 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290506 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290554 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290585 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290610 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290643 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290666 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290693 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290717 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290743 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290767 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290793 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290820 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290849 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290875 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290899 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290927 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290953 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290977 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291007 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291036 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291066 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291093 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291128 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291160 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291188 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291215 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291243 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291279 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291306 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291330 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291356 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291382 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291408 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291433 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291458 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291483 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291511 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291595 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291640 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291672 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292558 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292598 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292628 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292652 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292682 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292707 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292739 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292763 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292795 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292947 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292973 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293007 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293035 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293057 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293079 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293100 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293127 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293152 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293197 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293220 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293246 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293284 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293316 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293749 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293783 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293810 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293889 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-etc-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294293 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-multus-certs\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294392 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5f04b76-b225-449c-adc6-e979f26ab0da-mcd-auth-proxy-config\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294447 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfjd9\" (UniqueName: \"kubernetes.io/projected/d5f04b76-b225-449c-adc6-e979f26ab0da-kube-api-access-zfjd9\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294510 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/03885de7-3b56-436e-809f-f1ce7c027965-hosts-file\") pod \"node-resolver-j6gm8\" (UID: \"03885de7-3b56-436e-809f-f1ce7c027965\") " pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294563 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294594 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/271a556b-657c-4069-8cbc-a091395ca2ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294621 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294649 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d5f04b76-b225-449c-adc6-e979f26ab0da-rootfs\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294670 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5f04b76-b225-449c-adc6-e979f26ab0da-proxy-tls\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294691 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-bin\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294714 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlfx7\" (UniqueName: \"kubernetes.io/projected/ffeef997-af36-4315-8a12-46ecf69976e6-kube-api-access-tlfx7\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294738 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-cni-bin\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294762 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-netns\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294792 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294817 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-systemd\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294839 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294868 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294891 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-cni-multus\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294914 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294943 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-slash\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294975 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-socket-dir-parent\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295006 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295030 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-ovn\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295062 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295091 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295118 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295143 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdsv6\" (UniqueName: \"kubernetes.io/projected/03885de7-3b56-436e-809f-f1ce7c027965-kube-api-access-qdsv6\") pod \"node-resolver-j6gm8\" (UID: \"03885de7-3b56-436e-809f-f1ce7c027965\") " pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295166 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-systemd-units\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290317 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295250 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290431 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290439 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290521 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290606 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290783 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.290974 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291403 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291435 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.291712 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292320 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292741 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.292923 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293055 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293111 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293486 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293542 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293562 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293899 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.293925 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294165 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294263 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294296 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294394 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294776 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294781 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294922 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295016 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.294951 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295120 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295287 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295428 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.295721 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:27:55.795700886 +0000 UTC m=+41.152044911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.299388 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.299492 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.299631 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.299841 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.307941 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.300510 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.296242 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.296450 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.296465 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.296465 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.308816 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.296720 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.296936 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297312 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297331 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297370 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297590 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297761 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297942 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297868 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.297983 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.298009 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.298367 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.298386 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.298396 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.298501 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.298622 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.298737 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.298982 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.310563 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:55.810518381 +0000 UTC m=+41.166862406 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.310591 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.310649 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.310670 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.308548 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.310295 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.300575 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.310873 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.311056 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.308256 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.311965 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.312209 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.312414 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.312508 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.312617 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.312643 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.312930 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.312982 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.313311 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.313392 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.313460 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.313852 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.313860 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.314077 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.314093 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.314099 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295806 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.315044 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.315148 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.315880 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.316026 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.316128 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.316269 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.316858 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.295190 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcbgk\" (UniqueName: \"kubernetes.io/projected/271a556b-657c-4069-8cbc-a091395ca2ae-kube-api-access-fcbgk\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.316870 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.316975 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.317196 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.316279 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.317730 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.317785 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.317789 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.318116 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.318113 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.318408 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.318741 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.318933 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.319011 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.319029 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.319053 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.319566 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.319825 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.319860 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.319954 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.319978 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.320006 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.320021 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.320109 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.320357 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.320383 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.320391 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.320399 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.320488 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.320836 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.320935 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.321105 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.321307 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.321759 5035 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.322045 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.322095 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.322440 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.322865 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.322950 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.323098 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.323256 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.323479 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.323501 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.323754 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.323879 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324023 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324076 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324104 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324180 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324289 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324393 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324851 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.324203 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325127 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325740 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325816 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-cni-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325846 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-etc-kubernetes\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325872 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-script-lib\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325894 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-system-cni-dir\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325912 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-hostroot\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325934 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-daemon-config\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.325980 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-var-lib-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.326013 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.326035 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-node-log\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.326059 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5fafe5e6-83b2-447b-9379-b26e8071166b-cni-binary-copy\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.326082 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfvf9\" (UniqueName: \"kubernetes.io/projected/5fafe5e6-83b2-447b-9379-b26e8071166b-kube-api-access-qfvf9\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.326084 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.326103 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-netns\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.326124 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/271a556b-657c-4069-8cbc-a091395ca2ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.326812 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.328608 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.328713 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.329806 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.330795 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.330908 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.331001 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.331138 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.331152 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.331451 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.331766 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.334095 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:55.834069471 +0000 UTC m=+41.190413496 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.334234 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:55.834215946 +0000 UTC m=+41.190559971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.334264 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:55.834254887 +0000 UTC m=+41.190598912 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334520 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-env-overrides\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334583 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffeef997-af36-4315-8a12-46ecf69976e6-ovn-node-metrics-cert\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334611 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334636 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-kubelet\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334654 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-netd\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334671 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334691 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.334711 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-os-release\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.335305 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-system-cni-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.336732 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337461 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-os-release\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337565 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-log-socket\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337623 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-cnibin\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337656 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337682 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-kubelet\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337705 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337731 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337757 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-k8s-cni-cncf-io\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337789 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337812 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-config\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337836 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-cnibin\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337859 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-conf-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.337871 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338014 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338035 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338050 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338063 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338075 5035 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338087 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338099 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338113 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338125 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338137 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338149 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338160 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338175 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338188 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338199 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338210 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338222 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338233 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338244 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338255 5035 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338266 5035 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338277 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338288 5035 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338298 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338310 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338321 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338332 5035 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338464 5035 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338481 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338492 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338542 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338556 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338568 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338579 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338591 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338603 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338614 5035 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338624 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338637 5035 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338648 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338660 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338717 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338730 5035 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338744 5035 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.338757 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339014 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339032 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339521 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339046 5035 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339642 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339659 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339683 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339704 5035 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339695 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339719 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339808 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339822 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339832 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339843 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339853 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339861 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339870 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339879 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339888 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339897 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339906 5035 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339915 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339924 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339934 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339942 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340077 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340090 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340103 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340113 5035 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340121 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340130 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340139 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340148 5035 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340157 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340166 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340176 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339958 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.339955 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340094 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340185 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340309 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340323 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340352 5035 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340354 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340456 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.340468 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342599 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342623 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342637 5035 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342651 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342667 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342679 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342691 5035 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342703 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342716 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342712 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342727 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342783 5035 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342797 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342810 5035 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342825 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342839 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342868 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342883 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342896 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342908 5035 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342920 5035 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342931 5035 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342943 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342955 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342966 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342977 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.342990 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.343001 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.343013 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.343024 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.343036 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344160 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344179 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344190 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.343453 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.343133 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344202 5035 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344252 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344266 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344278 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344288 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344300 5035 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344311 5035 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344322 5035 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344334 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344346 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344357 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344366 5035 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344375 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344383 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344391 5035 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344401 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344411 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344419 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344428 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344436 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344445 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344455 5035 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344476 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344491 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344503 5035 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344514 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.343730 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344118 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344525 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344615 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344626 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344635 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344644 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344653 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344664 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344673 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344682 5035 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344692 5035 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344700 5035 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344709 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344717 5035 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344726 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344735 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344744 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344753 5035 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344761 5035 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344772 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344782 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344790 5035 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344799 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344807 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344816 5035 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344825 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344834 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344846 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344855 5035 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344864 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344873 5035 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344882 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344892 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344902 5035 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344913 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344923 5035 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344932 5035 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344941 5035 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.344950 5035 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.345512 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.354679 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.363190 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.363639 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.368696 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.369311 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.374089 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446205 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5fafe5e6-83b2-447b-9379-b26e8071166b-cni-binary-copy\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446250 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfvf9\" (UniqueName: \"kubernetes.io/projected/5fafe5e6-83b2-447b-9379-b26e8071166b-kube-api-access-qfvf9\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446267 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-netns\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446286 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-node-log\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446302 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-env-overrides\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446316 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffeef997-af36-4315-8a12-46ecf69976e6-ovn-node-metrics-cert\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446331 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/271a556b-657c-4069-8cbc-a091395ca2ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446345 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-kubelet\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446359 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-netd\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446376 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-os-release\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446395 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-system-cni-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446412 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-os-release\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446435 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446450 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-cnibin\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446466 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446479 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-log-socket\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446494 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446509 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-k8s-cni-cncf-io\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446523 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-kubelet\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446571 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446594 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-cnibin\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446612 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-conf-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446629 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-config\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446646 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5f04b76-b225-449c-adc6-e979f26ab0da-mcd-auth-proxy-config\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446664 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfjd9\" (UniqueName: \"kubernetes.io/projected/d5f04b76-b225-449c-adc6-e979f26ab0da-kube-api-access-zfjd9\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446683 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/03885de7-3b56-436e-809f-f1ce7c027965-hosts-file\") pod \"node-resolver-j6gm8\" (UID: \"03885de7-3b56-436e-809f-f1ce7c027965\") " pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446700 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-etc-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446717 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-multus-certs\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446734 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/271a556b-657c-4069-8cbc-a091395ca2ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446760 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d5f04b76-b225-449c-adc6-e979f26ab0da-rootfs\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446804 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-cni-bin\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446819 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5f04b76-b225-449c-adc6-e979f26ab0da-proxy-tls\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446837 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-bin\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446850 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlfx7\" (UniqueName: \"kubernetes.io/projected/ffeef997-af36-4315-8a12-46ecf69976e6-kube-api-access-tlfx7\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446874 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-netns\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446888 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-cni-multus\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446979 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-slash\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.446993 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-systemd\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447010 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447031 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447045 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-ovn\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447059 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-socket-dir-parent\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447079 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcbgk\" (UniqueName: \"kubernetes.io/projected/271a556b-657c-4069-8cbc-a091395ca2ae-kube-api-access-fcbgk\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447093 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-cni-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447109 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-etc-kubernetes\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447131 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdsv6\" (UniqueName: \"kubernetes.io/projected/03885de7-3b56-436e-809f-f1ce7c027965-kube-api-access-qdsv6\") pod \"node-resolver-j6gm8\" (UID: \"03885de7-3b56-436e-809f-f1ce7c027965\") " pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447164 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-systemd-units\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447179 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-script-lib\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447194 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-hostroot\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447244 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-daemon-config\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447290 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-var-lib-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447299 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-os-release\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447365 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-netns\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447442 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-slash\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447485 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-k8s-cni-cncf-io\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447500 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-cni-bin\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447510 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-ovn-kubernetes\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447562 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-cnibin\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447900 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447931 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-log-socket\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448005 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-cni-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448032 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-conf-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448059 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-cnibin\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448079 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-kubelet\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448095 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448105 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5fafe5e6-83b2-447b-9379-b26e8071166b-cni-binary-copy\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448108 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-node-log\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448132 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-os-release\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448141 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-cni-multus\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448151 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-netns\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448186 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/271a556b-657c-4069-8cbc-a091395ca2ae-cni-binary-copy\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-system-cni-dir\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448191 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d5f04b76-b225-449c-adc6-e979f26ab0da-mcd-auth-proxy-config\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448240 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-system-cni-dir\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.447455 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-netd\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448260 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-system-cni-dir\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448265 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448282 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-var-lib-kubelet\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448302 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-host-run-multus-certs\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448422 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d5f04b76-b225-449c-adc6-e979f26ab0da-rootfs\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448448 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-socket-dir-parent\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448462 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-etc-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448452 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/03885de7-3b56-436e-809f-f1ce7c027965-hosts-file\") pod \"node-resolver-j6gm8\" (UID: \"03885de7-3b56-436e-809f-f1ce7c027965\") " pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448484 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-etc-kubernetes\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448495 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-systemd-units\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448496 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-systemd\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448506 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-ovn\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448649 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-config\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.448687 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5fafe5e6-83b2-447b-9379-b26e8071166b-hostroot\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449105 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/271a556b-657c-4069-8cbc-a091395ca2ae-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449132 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5fafe5e6-83b2-447b-9379-b26e8071166b-multus-daemon-config\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449141 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449210 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-script-lib\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449250 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-var-lib-openvswitch\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449356 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-bin\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449628 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449653 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449668 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449683 5035 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449694 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449934 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449970 5035 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.449991 5035 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450006 5035 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450020 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450035 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450046 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450060 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450073 5035 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450086 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450207 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/271a556b-657c-4069-8cbc-a091395ca2ae-tuning-conf-dir\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.450306 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-env-overrides\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.454072 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffeef997-af36-4315-8a12-46ecf69976e6-ovn-node-metrics-cert\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.454097 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d5f04b76-b225-449c-adc6-e979f26ab0da-proxy-tls\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.467585 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcbgk\" (UniqueName: \"kubernetes.io/projected/271a556b-657c-4069-8cbc-a091395ca2ae-kube-api-access-fcbgk\") pod \"multus-additional-cni-plugins-4hwst\" (UID: \"271a556b-657c-4069-8cbc-a091395ca2ae\") " pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.467585 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfvf9\" (UniqueName: \"kubernetes.io/projected/5fafe5e6-83b2-447b-9379-b26e8071166b-kube-api-access-qfvf9\") pod \"multus-8rpfb\" (UID: \"5fafe5e6-83b2-447b-9379-b26e8071166b\") " pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.468285 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdsv6\" (UniqueName: \"kubernetes.io/projected/03885de7-3b56-436e-809f-f1ce7c027965-kube-api-access-qdsv6\") pod \"node-resolver-j6gm8\" (UID: \"03885de7-3b56-436e-809f-f1ce7c027965\") " pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.469292 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlfx7\" (UniqueName: \"kubernetes.io/projected/ffeef997-af36-4315-8a12-46ecf69976e6-kube-api-access-tlfx7\") pod \"ovnkube-node-cmx5n\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.470613 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfjd9\" (UniqueName: \"kubernetes.io/projected/d5f04b76-b225-449c-adc6-e979f26ab0da-kube-api-access-zfjd9\") pod \"machine-config-daemon-k6b5v\" (UID: \"d5f04b76-b225-449c-adc6-e979f26ab0da\") " pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.507614 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.516749 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.523178 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 09:27:55 crc kubenswrapper[5035]: W1002 09:27:55.524554 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-5576dad224955f939ceab6386ff2b50914a00912d64cd622cdc55ba18e496e9a WatchSource:0}: Error finding container 5576dad224955f939ceab6386ff2b50914a00912d64cd622cdc55ba18e496e9a: Status 404 returned error can't find the container with id 5576dad224955f939ceab6386ff2b50914a00912d64cd622cdc55ba18e496e9a Oct 02 09:27:55 crc kubenswrapper[5035]: W1002 09:27:55.529131 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-0537e72f6dc0cba019b42a572f7f28d64df2d3348f42306edc59936e84ac2880 WatchSource:0}: Error finding container 0537e72f6dc0cba019b42a572f7f28d64df2d3348f42306edc59936e84ac2880: Status 404 returned error can't find the container with id 0537e72f6dc0cba019b42a572f7f28d64df2d3348f42306edc59936e84ac2880 Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.529908 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-4hwst" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.536563 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:27:55 crc kubenswrapper[5035]: W1002 09:27:55.537061 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-6ec0fc636a2adf1dac8a16dea88923d46ace1cd06f8efae79b12f104948fec1a WatchSource:0}: Error finding container 6ec0fc636a2adf1dac8a16dea88923d46ace1cd06f8efae79b12f104948fec1a: Status 404 returned error can't find the container with id 6ec0fc636a2adf1dac8a16dea88923d46ace1cd06f8efae79b12f104948fec1a Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.543817 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-j6gm8" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.550446 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8rpfb" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.556274 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:27:55 crc kubenswrapper[5035]: W1002 09:27:55.591228 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5f04b76_b225_449c_adc6_e979f26ab0da.slice/crio-3ba4d5e91075d8cbc6ce2890fe54c04d17c345fd28e9b008cae8fae30bcc0ef5 WatchSource:0}: Error finding container 3ba4d5e91075d8cbc6ce2890fe54c04d17c345fd28e9b008cae8fae30bcc0ef5: Status 404 returned error can't find the container with id 3ba4d5e91075d8cbc6ce2890fe54c04d17c345fd28e9b008cae8fae30bcc0ef5 Oct 02 09:27:55 crc kubenswrapper[5035]: W1002 09:27:55.591667 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03885de7_3b56_436e_809f_f1ce7c027965.slice/crio-f3d363e2c1e5587a414ea5bed35279de1a4ef0300a5b03f6a20f6a8476accf17 WatchSource:0}: Error finding container f3d363e2c1e5587a414ea5bed35279de1a4ef0300a5b03f6a20f6a8476accf17: Status 404 returned error can't find the container with id f3d363e2c1e5587a414ea5bed35279de1a4ef0300a5b03f6a20f6a8476accf17 Oct 02 09:27:55 crc kubenswrapper[5035]: W1002 09:27:55.624973 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffeef997_af36_4315_8a12_46ecf69976e6.slice/crio-a19f9ec8b217f4387af40afcb84b5379cfb2c931ffcfa405d77acc72bf74787b WatchSource:0}: Error finding container a19f9ec8b217f4387af40afcb84b5379cfb2c931ffcfa405d77acc72bf74787b: Status 404 returned error can't find the container with id a19f9ec8b217f4387af40afcb84b5379cfb2c931ffcfa405d77acc72bf74787b Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.857330 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.857457 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:27:56.857434923 +0000 UTC m=+42.213778948 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.858451 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.858588 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.858696 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:55 crc kubenswrapper[5035]: I1002 09:27:55.858808 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.859008 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896295 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896342 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896356 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896404 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:56.85909934 +0000 UTC m=+42.215443365 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896433 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:56.896415216 +0000 UTC m=+42.252759241 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896482 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896524 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:56.896515019 +0000 UTC m=+42.252859044 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896674 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896712 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896727 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:55 crc kubenswrapper[5035]: E1002 09:27:55.896817 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:56.896794647 +0000 UTC m=+42.253138862 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.155480 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j6gm8" event={"ID":"03885de7-3b56-436e-809f-f1ce7c027965","Type":"ContainerStarted","Data":"f3d363e2c1e5587a414ea5bed35279de1a4ef0300a5b03f6a20f6a8476accf17"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.156702 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"3ba4d5e91075d8cbc6ce2890fe54c04d17c345fd28e9b008cae8fae30bcc0ef5"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.157362 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6ec0fc636a2adf1dac8a16dea88923d46ace1cd06f8efae79b12f104948fec1a"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.158028 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerStarted","Data":"3512d2e49898a4e8291db5f07686fcefd904ed0d577b85205f1b8f559d3a6cfd"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.158759 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0537e72f6dc0cba019b42a572f7f28d64df2d3348f42306edc59936e84ac2880"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.159402 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5576dad224955f939ceab6386ff2b50914a00912d64cd622cdc55ba18e496e9a"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.160469 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerStarted","Data":"6d18fa9f92901e921acd41f57ef0edb06ae7196cc4755d9072b7a24ef88411e4"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.161233 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"a19f9ec8b217f4387af40afcb84b5379cfb2c931ffcfa405d77acc72bf74787b"} Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.173110 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.173525 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.174335 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.175857 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.176615 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.177764 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.178315 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.178962 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.180061 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.180739 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.181704 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.182226 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.183616 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.184257 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.184840 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.185420 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.186848 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.187506 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.188704 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.189170 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.189814 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.190990 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.191463 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.193127 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.193894 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.195335 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.195834 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.196439 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.197307 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.198854 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.199478 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.201042 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.201720 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.202824 5035 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.202956 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.205430 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.205905 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.206386 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.206991 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.208801 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.209509 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.210436 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.211148 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.212350 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.213007 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.213690 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.214256 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.215036 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.216095 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.216761 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.217834 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.218467 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.219810 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.220315 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.221405 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.221651 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.221910 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.223042 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.223766 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.224303 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.230081 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.239488 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.246243 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.259070 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.267788 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.870037 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.870144 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.870286 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.870384 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:27:58.870345089 +0000 UTC m=+44.226689154 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.870447 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:58.870432541 +0000 UTC m=+44.226776566 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.970969 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.971048 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:27:56 crc kubenswrapper[5035]: I1002 09:27:56.971074 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971094 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971156 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:58.971142124 +0000 UTC m=+44.327486149 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971188 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971205 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971217 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971248 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:58.971238217 +0000 UTC m=+44.327582242 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971303 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971313 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971321 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:56 crc kubenswrapper[5035]: E1002 09:27:56.971346 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:27:58.97133805 +0000 UTC m=+44.327682075 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.162413 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.162447 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.162431 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:27:57 crc kubenswrapper[5035]: E1002 09:27:57.162600 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:27:57 crc kubenswrapper[5035]: E1002 09:27:57.162764 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:27:57 crc kubenswrapper[5035]: E1002 09:27:57.162848 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.165800 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b"} Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.166949 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerStarted","Data":"d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e"} Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.168939 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.169936 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-j6gm8" event={"ID":"03885de7-3b56-436e-809f-f1ce7c027965","Type":"ContainerStarted","Data":"1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886"} Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.171134 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193"} Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.172165 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerStarted","Data":"d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32"} Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.173176 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde"} Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.182923 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.193171 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.202758 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.210668 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.220161 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.234970 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.246189 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.263598 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.276111 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.284382 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:57 crc kubenswrapper[5035]: I1002 09:27:57.299117 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.177926 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99" exitCode=0 Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.178147 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.191135 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.209353 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.222848 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.249069 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.263235 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.275167 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.294589 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.307087 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.319793 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.331242 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.341675 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.353508 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.364418 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.373713 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.388068 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.399018 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.408728 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.423027 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.435957 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.448412 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.459709 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.471645 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.890069 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.890229 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.890329 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:28:02.890292194 +0000 UTC m=+48.246636259 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.890341 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.890411 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:02.890402077 +0000 UTC m=+48.246746302 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.991242 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.991293 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:27:58 crc kubenswrapper[5035]: I1002 09:27:58.991316 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991412 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991469 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:02.991449269 +0000 UTC m=+48.347793314 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991478 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991523 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991556 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991593 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991654 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991611 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:02.991593333 +0000 UTC m=+48.347937358 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991670 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:58 crc kubenswrapper[5035]: E1002 09:27:58.991740 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:02.991719407 +0000 UTC m=+48.348063432 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.162000 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.162095 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:27:59 crc kubenswrapper[5035]: E1002 09:27:59.162162 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:27:59 crc kubenswrapper[5035]: E1002 09:27:59.162268 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.162326 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:27:59 crc kubenswrapper[5035]: E1002 09:27:59.162403 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.183314 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb"} Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.187337 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.187385 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.189485 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d"} Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.191093 5035 generic.go:334] "Generic (PLEG): container finished" podID="271a556b-657c-4069-8cbc-a091395ca2ae" containerID="d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32" exitCode=0 Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.191164 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerDied","Data":"d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32"} Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.193850 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a"} Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.210772 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.225684 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.241260 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.254148 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.276276 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.290247 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.303482 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.315938 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.329513 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.343229 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.364121 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.381988 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.397102 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.412351 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.425221 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.438285 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.452749 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.471171 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.488000 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.502397 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.525059 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.539562 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.961178 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-8xzpk"] Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.962118 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.964570 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.964715 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.966979 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.967547 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.980602 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:27:59 crc kubenswrapper[5035]: I1002 09:27:59.995094 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:27:59Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.000653 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8222638-7aa2-4722-8013-a87cf909feae-serviceca\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.000762 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8222638-7aa2-4722-8013-a87cf909feae-host\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.000808 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5csb5\" (UniqueName: \"kubernetes.io/projected/e8222638-7aa2-4722-8013-a87cf909feae-kube-api-access-5csb5\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.011940 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.023962 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.040338 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.053144 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.071960 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.083753 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.101827 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8222638-7aa2-4722-8013-a87cf909feae-host\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.101720 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.101866 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5csb5\" (UniqueName: \"kubernetes.io/projected/e8222638-7aa2-4722-8013-a87cf909feae-kube-api-access-5csb5\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.101919 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8222638-7aa2-4722-8013-a87cf909feae-serviceca\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.101982 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8222638-7aa2-4722-8013-a87cf909feae-host\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.103095 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8222638-7aa2-4722-8013-a87cf909feae-serviceca\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.116601 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.134941 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5csb5\" (UniqueName: \"kubernetes.io/projected/e8222638-7aa2-4722-8013-a87cf909feae-kube-api-access-5csb5\") pod \"node-ca-8xzpk\" (UID: \"e8222638-7aa2-4722-8013-a87cf909feae\") " pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.135303 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.149458 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.200257 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.200330 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.200351 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.200364 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.202241 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerStarted","Data":"95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.217817 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.232331 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.245448 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.269973 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.283345 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.285091 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-8xzpk" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.300882 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: W1002 09:28:00.308886 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8222638_7aa2_4722_8013_a87cf909feae.slice/crio-b411b254b02491869fdcbf33d80dc12d1c27a7358ffccbeffe4f30d5c71ca428 WatchSource:0}: Error finding container b411b254b02491869fdcbf33d80dc12d1c27a7358ffccbeffe4f30d5c71ca428: Status 404 returned error can't find the container with id b411b254b02491869fdcbf33d80dc12d1c27a7358ffccbeffe4f30d5c71ca428 Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.321480 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.335686 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.350649 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.367140 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.382519 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.394577 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.574104 5035 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.576421 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.576462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.576476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.576620 5035 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.593682 5035 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.594188 5035 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.595762 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.595800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.595812 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.595831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.595843 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.609421 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.617478 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:28:00 crc kubenswrapper[5035]: E1002 09:28:00.627583 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.634691 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.635009 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.635244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.635333 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.635400 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.637383 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.642780 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 09:28:00 crc kubenswrapper[5035]: E1002 09:28:00.657901 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.661636 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.662952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.663122 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.663209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.663272 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.663325 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.682031 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: E1002 09:28:00.683312 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.692579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.693157 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.693246 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.693349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.693433 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.701945 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: E1002 09:28:00.707314 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.713182 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.713494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.713641 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.713734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.713820 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.719764 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: E1002 09:28:00.725981 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: E1002 09:28:00.726473 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.728435 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.728495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.728508 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.728553 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.728569 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.734968 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.750799 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.771458 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.783925 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.799839 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.815799 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.826853 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.830679 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.830703 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.830711 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.830724 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.830734 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.840185 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.851641 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.863460 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.879592 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.893578 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.906337 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.920059 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.932229 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.933503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.933613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.933625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.933643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.933658 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:00Z","lastTransitionTime":"2025-10-02T09:28:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.955780 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.967699 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.981186 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:00 crc kubenswrapper[5035]: I1002 09:28:00.992828 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:00Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.008593 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.037961 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.038270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.038283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.038346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.038362 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.141621 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.141675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.141709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.141726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.141736 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.162233 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:01 crc kubenswrapper[5035]: E1002 09:28:01.162432 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.162570 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.162570 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:01 crc kubenswrapper[5035]: E1002 09:28:01.162751 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:01 crc kubenswrapper[5035]: E1002 09:28:01.162885 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.208974 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8xzpk" event={"ID":"e8222638-7aa2-4722-8013-a87cf909feae","Type":"ContainerStarted","Data":"59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.209035 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-8xzpk" event={"ID":"e8222638-7aa2-4722-8013-a87cf909feae","Type":"ContainerStarted","Data":"b411b254b02491869fdcbf33d80dc12d1c27a7358ffccbeffe4f30d5c71ca428"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.211873 5035 generic.go:334] "Generic (PLEG): container finished" podID="271a556b-657c-4069-8cbc-a091395ca2ae" containerID="95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c" exitCode=0 Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.212006 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerDied","Data":"95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c"} Oct 02 09:28:01 crc kubenswrapper[5035]: E1002 09:28:01.219695 5035 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.232497 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.244205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.244257 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.244273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.244292 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.244303 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.247983 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.263278 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.282733 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.297317 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.315524 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.329365 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.345004 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.346748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.346781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.346794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.346811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.346824 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.361275 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.377383 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.390552 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.402406 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.414605 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.425435 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.435766 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.448886 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.451253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.451285 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.451296 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.451310 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.451347 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.462274 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.477311 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.489624 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.500848 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.509903 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.529608 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.541311 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.554391 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.554441 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.554453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.554472 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.554486 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.557137 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.569522 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.587771 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:01Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.657357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.657401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.657410 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.657427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.657438 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.760663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.760707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.760718 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.760735 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.760747 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.863967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.864014 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.864024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.864040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.864051 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.966713 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.966746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.966754 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.966767 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:01 crc kubenswrapper[5035]: I1002 09:28:01.966776 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:01Z","lastTransitionTime":"2025-10-02T09:28:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.069565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.069601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.069613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.069629 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.069642 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.171357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.171397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.171411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.171424 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.171434 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.217706 5035 generic.go:334] "Generic (PLEG): container finished" podID="271a556b-657c-4069-8cbc-a091395ca2ae" containerID="d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925" exitCode=0 Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.217769 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerDied","Data":"d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.233954 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.248034 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.264025 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.277746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.278299 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.278318 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.278332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.278343 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.281196 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.300909 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.314651 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.330284 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.345364 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.360605 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.371956 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.381668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.381730 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.381741 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.381784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.381803 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.385197 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.401312 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.415760 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.483927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.483962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.483970 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.483983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.483993 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.586744 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.586797 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.586809 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.586827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.586842 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.660885 5035 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.689628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.689674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.689685 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.689705 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.689718 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.792929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.792986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.792997 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.793016 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.793030 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.895493 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.895559 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.895571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.895588 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.895598 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.932155 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:28:02 crc kubenswrapper[5035]: E1002 09:28:02.932346 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:28:10.932319104 +0000 UTC m=+56.288663129 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.932402 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:02 crc kubenswrapper[5035]: E1002 09:28:02.932511 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:02 crc kubenswrapper[5035]: E1002 09:28:02.932588 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:10.932574732 +0000 UTC m=+56.288918757 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.997972 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.998016 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.998031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.998048 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:02 crc kubenswrapper[5035]: I1002 09:28:02.998059 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:02Z","lastTransitionTime":"2025-10-02T09:28:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.033728 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.033793 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.033837 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.033925 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.033937 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.033962 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.033975 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.033994 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:11.033978394 +0000 UTC m=+56.390322429 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.034014 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:11.034002685 +0000 UTC m=+56.390346710 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.034042 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.034086 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.034100 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.034173 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:11.034152769 +0000 UTC m=+56.390496954 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.100006 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.100055 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.100064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.100083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.100093 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.162931 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.163056 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.163079 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.163243 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.163273 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.163309 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.202345 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.202426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.202448 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.202475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.202493 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.304936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.304979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.304991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.305009 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.305021 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.407420 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.407481 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.407494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.407514 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.407547 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.509262 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.509293 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.509301 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.509317 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.509331 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.612459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.612500 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.612512 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.612545 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.612561 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.715792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.715850 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.715864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.715890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.715905 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.819195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.819254 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.819265 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.819282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.819294 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.861382 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.876244 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.876814 5035 scope.go:117] "RemoveContainer" containerID="1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b" Oct 02 09:28:03 crc kubenswrapper[5035]: E1002 09:28:03.877175 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.878552 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.891752 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.905075 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.917104 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.921378 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.921417 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.921429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.921445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.921458 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:03Z","lastTransitionTime":"2025-10-02T09:28:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.929252 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.942054 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.953292 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.965997 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.978628 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:03 crc kubenswrapper[5035]: I1002 09:28:03.998458 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:03Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.007630 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.023680 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.024361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.024388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.024398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.024413 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.024423 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.032907 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.126992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.127027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.127037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.127052 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.127063 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.227457 5035 generic.go:334] "Generic (PLEG): container finished" podID="271a556b-657c-4069-8cbc-a091395ca2ae" containerID="fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7" exitCode=0 Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.227521 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerDied","Data":"fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.228487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.228519 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.228553 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.228576 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.228588 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.234885 5035 scope.go:117] "RemoveContainer" containerID="1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.235352 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.244038 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.258079 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.272369 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.284585 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.296823 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.310326 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.323767 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.331426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.331478 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.331494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.331513 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.331540 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.337868 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.350609 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.371851 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.383634 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.399383 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.417754 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.432337 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:04Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.433687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.433713 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.433722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.433736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.433745 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.535908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.536137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.536245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.536361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.536453 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.639403 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.639470 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.639482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.639505 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.639518 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.741965 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.742308 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.742387 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.742463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.742592 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.845059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.845100 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.845111 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.845127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.845140 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.948411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.948455 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.948467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.948483 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:04 crc kubenswrapper[5035]: I1002 09:28:04.948493 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:04Z","lastTransitionTime":"2025-10-02T09:28:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.051287 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.051330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.051342 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.051358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.051371 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.153798 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.153846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.153859 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.153878 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.153889 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.162378 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.162463 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.162410 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:05 crc kubenswrapper[5035]: E1002 09:28:05.162614 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:05 crc kubenswrapper[5035]: E1002 09:28:05.162723 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:05 crc kubenswrapper[5035]: E1002 09:28:05.162801 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.240991 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerStarted","Data":"a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.243171 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.245344 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.245767 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.256973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.257025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.257036 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.257053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.257064 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.257798 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.270990 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.284931 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.300783 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.314888 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.326957 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.340090 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.355214 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.359819 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.359853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.359867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.359883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.359892 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.369489 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.386258 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.396204 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.415925 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.425309 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.438617 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.457138 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.462132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.462176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.462190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.462208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.462221 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.470234 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.484262 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.498708 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.512404 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.526365 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.540406 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.553886 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.566001 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.566072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.566087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.566107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.566120 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.568064 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.583891 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.597771 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.611437 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.624522 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.637945 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:05Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.668694 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.668727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.668736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.668751 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.668761 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.771420 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.771473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.771487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.771507 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.771520 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.874625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.875005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.875024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.875050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.875064 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.976969 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.977012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.977023 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.977039 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:05 crc kubenswrapper[5035]: I1002 09:28:05.977052 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:05Z","lastTransitionTime":"2025-10-02T09:28:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.079493 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.079542 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.079552 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.079568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.079579 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.182574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.182617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.182628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.182645 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.182657 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.182778 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.193114 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.207593 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.220604 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.234345 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.248967 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.256508 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.257306 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.257351 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.257366 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.261202 5035 generic.go:334] "Generic (PLEG): container finished" podID="271a556b-657c-4069-8cbc-a091395ca2ae" containerID="a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8" exitCode=0 Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.261295 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerDied","Data":"a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.263801 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.279295 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.285836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.285865 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.285873 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.285887 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.285896 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.293015 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.301907 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.302072 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.317463 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.333023 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.350487 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.362243 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.373149 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.385266 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.392416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.392462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.392475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.392492 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.392503 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.397503 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.408120 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.425128 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.436612 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.451086 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.463943 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.475613 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.488484 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.495359 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.495399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.495412 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.495427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.495437 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.501625 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.513752 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.527569 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.541331 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.553630 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.572431 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7"] Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.573033 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.575197 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.576470 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.585308 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.597297 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.598467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.599087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.599170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.599299 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.599412 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.610810 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.624835 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.647976 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.660736 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.671794 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88a34b1e-c1fe-435c-9cfa-3288d778139c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.671857 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kxwn\" (UniqueName: \"kubernetes.io/projected/88a34b1e-c1fe-435c-9cfa-3288d778139c-kube-api-access-7kxwn\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.671898 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88a34b1e-c1fe-435c-9cfa-3288d778139c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.672017 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88a34b1e-c1fe-435c-9cfa-3288d778139c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.677160 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.690560 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.703164 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.703205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.703216 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.703231 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.703243 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.703828 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.719125 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.735655 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.750326 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.764211 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.772868 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88a34b1e-c1fe-435c-9cfa-3288d778139c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.772913 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kxwn\" (UniqueName: \"kubernetes.io/projected/88a34b1e-c1fe-435c-9cfa-3288d778139c-kube-api-access-7kxwn\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.772952 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88a34b1e-c1fe-435c-9cfa-3288d778139c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.772991 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88a34b1e-c1fe-435c-9cfa-3288d778139c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.773561 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/88a34b1e-c1fe-435c-9cfa-3288d778139c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.773793 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/88a34b1e-c1fe-435c-9cfa-3288d778139c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.778605 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.779630 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/88a34b1e-c1fe-435c-9cfa-3288d778139c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.789379 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kxwn\" (UniqueName: \"kubernetes.io/projected/88a34b1e-c1fe-435c-9cfa-3288d778139c-kube-api-access-7kxwn\") pod \"ovnkube-control-plane-749d76644c-5k5w7\" (UID: \"88a34b1e-c1fe-435c-9cfa-3288d778139c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.793415 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.806429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.806484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.806500 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.806518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.806554 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.889275 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.908974 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.909029 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.909040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.909057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:06 crc kubenswrapper[5035]: I1002 09:28:06.909109 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:06Z","lastTransitionTime":"2025-10-02T09:28:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.012661 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.012717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.012728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.012747 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.012759 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.115126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.115171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.115181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.115200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.115212 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.162469 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:07 crc kubenswrapper[5035]: E1002 09:28:07.162644 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.163096 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:07 crc kubenswrapper[5035]: E1002 09:28:07.163161 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.163217 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:07 crc kubenswrapper[5035]: E1002 09:28:07.163274 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.217857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.217905 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.217917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.217934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.217945 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.267636 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" event={"ID":"88a34b1e-c1fe-435c-9cfa-3288d778139c","Type":"ContainerStarted","Data":"7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.267711 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" event={"ID":"88a34b1e-c1fe-435c-9cfa-3288d778139c","Type":"ContainerStarted","Data":"cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.267727 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" event={"ID":"88a34b1e-c1fe-435c-9cfa-3288d778139c","Type":"ContainerStarted","Data":"4bea39d1b2b2c2206fc631e0f24f1d523723fe9c2a46fb1e8374cb3e8bc70a0f"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.271803 5035 generic.go:334] "Generic (PLEG): container finished" podID="271a556b-657c-4069-8cbc-a091395ca2ae" containerID="8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9" exitCode=0 Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.271876 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerDied","Data":"8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.286357 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.300065 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.320599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.320645 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.320664 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.320685 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.320697 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.322319 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.335477 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.347741 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.363034 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.376830 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.390061 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.406400 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.421420 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.423363 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.423396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.423415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.423475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.423489 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.434012 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.445781 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.458547 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.470615 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.484317 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.526798 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.526836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.526847 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.526865 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.526877 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.629092 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.629151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.629167 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.629185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.629199 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.683732 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hzsjk"] Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.684176 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:07 crc kubenswrapper[5035]: E1002 09:28:07.684233 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.702392 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.713462 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.726073 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.731382 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.731429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.731439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.731457 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.731470 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.740967 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.755453 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.768221 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.780756 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.783525 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.783598 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2vpf\" (UniqueName: \"kubernetes.io/projected/452af00b-602d-43ab-a345-5453d6aebcf0-kube-api-access-c2vpf\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.795648 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.810746 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.825509 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.833860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.833909 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.833924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.833943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.833956 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.838031 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.859093 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.873338 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.884259 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.884359 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2vpf\" (UniqueName: \"kubernetes.io/projected/452af00b-602d-43ab-a345-5453d6aebcf0-kube-api-access-c2vpf\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:07 crc kubenswrapper[5035]: E1002 09:28:07.884513 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:07 crc kubenswrapper[5035]: E1002 09:28:07.884812 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:08.384795179 +0000 UTC m=+53.741139204 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.889513 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.904332 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2vpf\" (UniqueName: \"kubernetes.io/projected/452af00b-602d-43ab-a345-5453d6aebcf0-kube-api-access-c2vpf\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.908552 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.926435 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:07Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.936711 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.936778 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.936791 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.936815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:07 crc kubenswrapper[5035]: I1002 09:28:07.936830 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:07Z","lastTransitionTime":"2025-10-02T09:28:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.039903 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.039955 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.039967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.040029 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.040045 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.142105 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.142153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.142164 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.142183 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.142196 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.245121 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.245188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.245201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.245221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.245233 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.282040 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" event={"ID":"271a556b-657c-4069-8cbc-a091395ca2ae","Type":"ContainerStarted","Data":"dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.296642 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.310760 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.322993 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.337639 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.347280 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.347371 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.347383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.347402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.347416 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.349172 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.364291 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.376948 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.387773 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.390234 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:08 crc kubenswrapper[5035]: E1002 09:28:08.390639 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:08 crc kubenswrapper[5035]: E1002 09:28:08.390684 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:09.39067181 +0000 UTC m=+54.747015835 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.400899 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.413001 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.423768 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.438681 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.450379 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.450430 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.450447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.450467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.450478 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.452284 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.467313 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.479775 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.500278 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.515486 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.529132 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.542615 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.553125 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.553197 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.553208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.553225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.553235 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.561447 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.574252 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.594346 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.619745 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.640273 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.653777 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.655462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.655492 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.655502 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.655517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.655550 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.666461 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.678570 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.690851 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.705601 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.715351 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.732015 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.741472 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:08Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.757762 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.757807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.757820 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.757840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.757854 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.859616 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.859882 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.859948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.860020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.860087 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.962891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.962948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.962960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.962976 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:08 crc kubenswrapper[5035]: I1002 09:28:08.962986 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:08Z","lastTransitionTime":"2025-10-02T09:28:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.065608 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.065666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.065681 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.065702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.065715 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.162621 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.162654 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.162659 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.162641 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:09 crc kubenswrapper[5035]: E1002 09:28:09.162795 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:09 crc kubenswrapper[5035]: E1002 09:28:09.162889 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:09 crc kubenswrapper[5035]: E1002 09:28:09.162985 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:09 crc kubenswrapper[5035]: E1002 09:28:09.163049 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.168234 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.168300 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.168314 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.168333 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.168349 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.270870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.270924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.270934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.270958 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.270976 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.374150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.374229 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.374241 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.374259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.374270 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.397732 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:09 crc kubenswrapper[5035]: E1002 09:28:09.397870 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:09 crc kubenswrapper[5035]: E1002 09:28:09.397944 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:11.397924116 +0000 UTC m=+56.754268141 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.476357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.476719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.476816 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.476893 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.476949 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.579244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.579293 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.579306 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.579323 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.579335 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.681464 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.681671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.681681 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.681698 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.681708 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.783404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.783444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.783490 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.783506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.783516 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.885846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.885884 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.885897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.885915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.885929 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.988618 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.988657 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.988666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.988681 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:09 crc kubenswrapper[5035]: I1002 09:28:09.988690 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:09Z","lastTransitionTime":"2025-10-02T09:28:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.091721 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.091830 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.091851 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.091876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.091888 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.194445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.194480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.194490 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.194506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.194523 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.291761 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/0.log" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.295581 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0" exitCode=1 Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.295638 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.296579 5035 scope.go:117] "RemoveContainer" containerID="78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.298167 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.298200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.298211 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.298229 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.298241 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.314921 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.352802 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.370559 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.385615 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.400100 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.400744 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.400776 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.400787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.400804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.400815 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.414698 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.432031 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.449683 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.465176 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.479519 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.505570 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.505668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.505686 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.505710 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.505731 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.508656 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:10Z\\\",\\\"message\\\":\\\"de event handler 7 for removal\\\\nI1002 09:28:09.807620 6353 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:28:09.807641 6353 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:28:09.807713 6353 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.807718 6353 factory.go:656] Stopping watch factory\\\\nI1002 09:28:09.807743 6353 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:28:09.807765 6353 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:28:09.807776 6353 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:09.807788 6353 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 09:28:09.807799 6353 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:28:09.808424 6353 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.808638 6353 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.808870 6353 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.809062 6353 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.523269 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.539216 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.554076 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.571381 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.586903 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:10Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.609205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.609256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.609266 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.609280 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.609291 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.712076 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.712442 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.712524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.712632 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.712704 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.815077 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.815125 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.815137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.815153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.815167 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.917986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.918033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.918047 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.918067 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:10 crc kubenswrapper[5035]: I1002 09:28:10.918080 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:10Z","lastTransitionTime":"2025-10-02T09:28:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.015315 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.015491 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.015644 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.015694 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:27.015681407 +0000 UTC m=+72.372025422 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.015833 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:28:27.015802931 +0000 UTC m=+72.372146956 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.021856 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.021907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.021916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.021931 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.021943 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.090117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.090336 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.090409 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.090486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.090566 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.104821 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.109368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.109415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.109429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.109445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.109454 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.116194 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.116243 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.116277 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116392 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116408 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116459 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116421 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116473 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116480 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116484 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116562 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:27.116520674 +0000 UTC m=+72.472864789 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116587 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:27.116576676 +0000 UTC m=+72.472920901 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.116606 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:27.116597196 +0000 UTC m=+72.472941431 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.122827 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.127422 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.127506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.127520 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.127585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.127608 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.142063 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.146269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.146300 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.146311 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.146330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.146341 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.161653 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.162126 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.162162 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.162179 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.162150 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.162279 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.162341 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.162443 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.162544 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.165802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.165829 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.165841 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.165857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.165868 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.177254 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:11Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.177378 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.178831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.178859 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.178869 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.178885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.178907 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.281136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.281171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.281182 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.281197 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.281208 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.300600 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/0.log" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.303559 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.383973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.384019 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.384030 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.384046 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.384057 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.419641 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.419856 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: E1002 09:28:11.419915 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:15.419897969 +0000 UTC m=+60.776241994 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.486235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.486274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.486284 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.486298 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.486308 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.588997 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.589050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.589062 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.589080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.589093 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.691325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.691368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.691391 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.691405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.691415 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.793699 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.793736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.793748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.793763 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.793773 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.897088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.897143 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.897152 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.897170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.897179 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.999462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.999505 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.999513 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.999552 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:11 crc kubenswrapper[5035]: I1002 09:28:11.999564 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:11Z","lastTransitionTime":"2025-10-02T09:28:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.103336 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.103390 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.103403 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.103421 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.103433 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.204958 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.204996 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.205007 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.205023 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.205036 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.306732 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.307203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.307237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.307245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.307261 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.307272 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.321398 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.335882 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.350477 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.367178 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.380151 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.392387 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.405783 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.409076 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.409111 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.409121 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.409136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.409146 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.420013 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.432861 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.446347 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.462362 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.474165 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.491582 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.522081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.522164 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.522200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.522219 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.522230 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.526395 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.546750 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:10Z\\\",\\\"message\\\":\\\"de event handler 7 for removal\\\\nI1002 09:28:09.807620 6353 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:28:09.807641 6353 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:28:09.807713 6353 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.807718 6353 factory.go:656] Stopping watch factory\\\\nI1002 09:28:09.807743 6353 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:28:09.807765 6353 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:28:09.807776 6353 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:09.807788 6353 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 09:28:09.807799 6353 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:28:09.808424 6353 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.808638 6353 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.808870 6353 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.809062 6353 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.558409 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.624313 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.624356 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.624368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.624383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.624395 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.727611 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.727668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.727677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.727698 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.727710 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.831130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.831193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.831205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.831226 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.831239 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.934301 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.934359 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.934370 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.934385 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:12 crc kubenswrapper[5035]: I1002 09:28:12.934418 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:12Z","lastTransitionTime":"2025-10-02T09:28:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.036840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.036886 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.036897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.036913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.036926 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.139565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.139625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.139640 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.139662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.139677 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.162235 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.162269 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.162276 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:13 crc kubenswrapper[5035]: E1002 09:28:13.162391 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.162414 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:13 crc kubenswrapper[5035]: E1002 09:28:13.162631 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:13 crc kubenswrapper[5035]: E1002 09:28:13.162788 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:13 crc kubenswrapper[5035]: E1002 09:28:13.162915 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.242210 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.242461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.242612 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.242711 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.242782 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.311485 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/1.log" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.312500 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/0.log" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.315176 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f" exitCode=1 Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.315303 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.315431 5035 scope.go:117] "RemoveContainer" containerID="78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.316125 5035 scope.go:117] "RemoveContainer" containerID="1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f" Oct 02 09:28:13 crc kubenswrapper[5035]: E1002 09:28:13.316355 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.332760 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.344670 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.345091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.345144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.345166 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.345181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.345192 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.357668 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.368319 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.380268 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.391435 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.407648 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.419670 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.432062 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.442180 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.447051 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.447090 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.447123 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.447139 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.447151 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.459946 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78f74499da884f9a9e7705ba8afb6540051d8bff4fa17921f1e9ab31adbc35d0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:10Z\\\",\\\"message\\\":\\\"de event handler 7 for removal\\\\nI1002 09:28:09.807620 6353 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 09:28:09.807641 6353 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 09:28:09.807713 6353 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.807718 6353 factory.go:656] Stopping watch factory\\\\nI1002 09:28:09.807743 6353 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 09:28:09.807765 6353 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 09:28:09.807776 6353 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:09.807788 6353 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 09:28:09.807799 6353 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 09:28:09.808424 6353 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.808638 6353 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.808870 6353 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:09.809062 6353 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.468147 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.478109 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.489463 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.504371 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.516558 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:13Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.550241 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.550293 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.550309 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.550326 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.550338 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.653947 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.654405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.654509 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.654624 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.654718 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.757718 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.757760 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.757771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.757787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.757797 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.859975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.860037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.860051 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.860071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.860085 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.963344 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.963387 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.963397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.963416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:13 crc kubenswrapper[5035]: I1002 09:28:13.963430 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:13Z","lastTransitionTime":"2025-10-02T09:28:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.066242 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.066282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.066293 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.066308 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.066320 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.168286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.168332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.168341 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.168352 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.168362 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.270743 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.270788 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.270797 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.270810 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.270819 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.321137 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/1.log" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.324972 5035 scope.go:117] "RemoveContainer" containerID="1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f" Oct 02 09:28:14 crc kubenswrapper[5035]: E1002 09:28:14.325164 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.339079 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.360403 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.372047 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.373651 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.373710 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.373724 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.373746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.373757 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.384690 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.398720 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.419926 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.432287 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.446012 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.461120 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.477147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.477188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.477200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.477217 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.477231 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.488129 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.506597 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.522562 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.535761 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.552033 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.566148 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.579429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.579469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.579480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.579497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.579508 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.580640 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:14Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.682245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.682702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.682837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.682924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.682995 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.786336 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.786375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.786388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.786404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.786415 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.895748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.896004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.896158 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.896278 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.896389 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.998983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.999024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.999038 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.999060 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:14 crc kubenswrapper[5035]: I1002 09:28:14.999074 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:14Z","lastTransitionTime":"2025-10-02T09:28:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.003214 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.016042 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.032632 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.047321 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.060524 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.072261 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.089770 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.101825 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.102041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.102118 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.102191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.102264 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.102674 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.115337 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.130417 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.145619 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.159674 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.162201 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.162206 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.162326 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:15 crc kubenswrapper[5035]: E1002 09:28:15.162456 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:15 crc kubenswrapper[5035]: E1002 09:28:15.162506 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:15 crc kubenswrapper[5035]: E1002 09:28:15.162605 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.162610 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:15 crc kubenswrapper[5035]: E1002 09:28:15.162719 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.172657 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.186494 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.197760 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.204634 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.204872 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.204984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.205087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.205215 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.208297 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.221371 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:15Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.308236 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.308603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.308690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.308838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.308947 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.412172 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.412495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.412719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.412874 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.413004 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.463060 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:15 crc kubenswrapper[5035]: E1002 09:28:15.463581 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:15 crc kubenswrapper[5035]: E1002 09:28:15.463898 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:23.463880766 +0000 UTC m=+68.820224791 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.515762 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.515829 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.515847 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.515880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.515900 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.618459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.618499 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.618508 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.618523 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.618549 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.720489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.720524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.720564 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.720580 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.720591 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.823852 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.823930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.823956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.823982 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.823998 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.926628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.926693 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.926709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.926731 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:15 crc kubenswrapper[5035]: I1002 09:28:15.926744 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:15Z","lastTransitionTime":"2025-10-02T09:28:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.030082 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.030130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.030146 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.030178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.030196 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.132489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.132558 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.132574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.132592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.132607 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.181128 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.198130 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.216301 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.235255 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.235831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.235960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.235999 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.236028 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.237714 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.254983 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.273845 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.289646 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.301628 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.312686 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.323780 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.337926 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.338956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.339017 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.339042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.339072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.339096 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.351632 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.365968 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.378768 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.400992 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.413113 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:16Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.446259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.446305 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.446318 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.446336 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.446348 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.548465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.548506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.548517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.548557 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.548570 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.655870 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.655941 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.655967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.656000 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.656026 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.758068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.758112 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.758125 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.758141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.758155 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.861144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.861223 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.861247 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.861285 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.861311 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.965738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.965793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.965809 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.965827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:16 crc kubenswrapper[5035]: I1002 09:28:16.965840 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:16Z","lastTransitionTime":"2025-10-02T09:28:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.067763 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.067793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.067806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.067827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.067837 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.162382 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.162375 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.162483 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:17 crc kubenswrapper[5035]: E1002 09:28:17.162668 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.162769 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:17 crc kubenswrapper[5035]: E1002 09:28:17.162930 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:17 crc kubenswrapper[5035]: E1002 09:28:17.163067 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:17 crc kubenswrapper[5035]: E1002 09:28:17.163142 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.170946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.171018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.171043 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.171073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.171094 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.274168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.274748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.274766 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.274790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.274811 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.378571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.378632 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.378649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.378672 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.378696 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.481636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.481706 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.481726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.481753 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.481770 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.584412 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.584467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.584476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.584496 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.584505 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.687200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.687318 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.687330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.687357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.687371 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.789628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.789672 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.789687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.789705 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.789717 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.892297 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.892336 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.892345 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.892357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.892366 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.994017 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.994057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.994066 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.994078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:17 crc kubenswrapper[5035]: I1002 09:28:17.994087 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:17Z","lastTransitionTime":"2025-10-02T09:28:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.096253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.096289 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.096300 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.096312 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.096321 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.198281 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.198325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.198335 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.198350 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.198358 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.300519 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.300731 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.300812 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.300842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.300867 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.403019 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.403080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.403097 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.403114 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.403126 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.505659 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.505717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.505733 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.505757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.505773 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.608248 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.608286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.608295 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.608307 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.608316 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.710606 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.710656 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.710668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.710686 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.710702 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.814319 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.814356 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.814367 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.814384 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.814394 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.916666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.916714 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.916722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.916737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:18 crc kubenswrapper[5035]: I1002 09:28:18.916747 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:18Z","lastTransitionTime":"2025-10-02T09:28:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.019828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.019907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.019929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.019956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.019976 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.123501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.123656 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.123675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.123700 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.123720 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.162250 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.162359 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:19 crc kubenswrapper[5035]: E1002 09:28:19.162511 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.162578 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:19 crc kubenswrapper[5035]: E1002 09:28:19.162700 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.162712 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:19 crc kubenswrapper[5035]: E1002 09:28:19.162891 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:19 crc kubenswrapper[5035]: E1002 09:28:19.163049 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.227154 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.227209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.227225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.227250 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.227267 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.331297 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.331359 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.331376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.331399 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.331415 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.434016 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.434063 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.434075 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.434094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.434107 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.536633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.536709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.536737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.536767 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.536790 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.639907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.639950 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.639963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.639978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.639988 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.743099 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.743144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.743155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.743170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.743179 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.846911 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.847002 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.847040 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.847073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.847096 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.949765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.949799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.949811 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.949829 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:19 crc kubenswrapper[5035]: I1002 09:28:19.949841 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:19Z","lastTransitionTime":"2025-10-02T09:28:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.052745 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.052874 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.052900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.052925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.052942 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.156829 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.156888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.156905 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.156931 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.156948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.260415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.260467 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.260479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.260500 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.260516 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.363091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.363161 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.363179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.363204 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.363222 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.465804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.465834 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.465842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.465855 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.465865 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.568269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.568307 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.568316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.568332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.568342 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.670509 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.670562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.670571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.670585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.670595 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.773118 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.773171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.773179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.773203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.773216 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.876369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.876433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.876442 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.876454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.876465 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.980002 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.980072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.980080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.980094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:20 crc kubenswrapper[5035]: I1002 09:28:20.980103 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:20Z","lastTransitionTime":"2025-10-02T09:28:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.083086 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.083133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.083146 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.083163 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.083174 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.161943 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.162019 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.161964 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.162127 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.162191 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.162299 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.162393 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.162509 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.187806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.187910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.187934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.188004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.188027 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.290920 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.291032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.291068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.291096 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.291115 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.322349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.322453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.322484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.322580 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.322746 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.345661 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.350764 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.350991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.351058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.351129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.351201 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.374405 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.380478 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.380771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.380840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.380920 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.380990 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.396424 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.400251 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.400316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.400332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.400361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.400378 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.420915 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.425779 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.425953 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.426026 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.426089 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.426427 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.445894 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: E1002 09:28:21.446088 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.447804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.447879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.447902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.447931 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.447955 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.551284 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.551414 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.551488 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.551518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.551581 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.569736 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.581746 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.588137 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.604558 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.636415 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.648610 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.657018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.657062 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.657073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.657087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.657101 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.665456 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.679109 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.690799 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.700929 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.712822 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.723758 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.734182 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.746062 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.756126 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.760415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.760444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.760454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.760469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.760481 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.781259 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.824630 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.836724 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:21Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.864390 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.865119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.865167 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.865195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.865212 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.967899 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.967960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.967977 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.968002 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:21 crc kubenswrapper[5035]: I1002 09:28:21.968019 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:21Z","lastTransitionTime":"2025-10-02T09:28:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.070910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.070981 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.071006 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.071034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.071056 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.173063 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.173101 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.173111 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.173124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.173132 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.275492 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.275541 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.275554 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.275570 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.275583 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.379008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.379068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.379080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.379106 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.379127 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.482168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.482216 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.482230 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.482248 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.482260 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.584204 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.584235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.584247 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.584260 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.584269 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.687251 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.687321 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.687341 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.687367 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.687384 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.790342 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.790396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.790411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.790433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.790448 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.892841 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.892881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.892890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.892908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.892917 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.995883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.995914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.995924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.995938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:22 crc kubenswrapper[5035]: I1002 09:28:22.995950 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:22Z","lastTransitionTime":"2025-10-02T09:28:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.098663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.098720 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.098738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.098763 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.098784 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.162480 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.162578 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.162623 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.162647 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:23 crc kubenswrapper[5035]: E1002 09:28:23.162812 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:23 crc kubenswrapper[5035]: E1002 09:28:23.162959 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:23 crc kubenswrapper[5035]: E1002 09:28:23.163124 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:23 crc kubenswrapper[5035]: E1002 09:28:23.163286 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.201432 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.201496 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.201515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.201569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.201586 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.303907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.303936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.303945 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.303959 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.303969 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.406380 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.406420 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.406431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.406444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.406455 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.465187 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:23 crc kubenswrapper[5035]: E1002 09:28:23.465350 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:23 crc kubenswrapper[5035]: E1002 09:28:23.465413 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:39.465396334 +0000 UTC m=+84.821740359 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.508840 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.508876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.508885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.508897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.508905 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.611635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.611679 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.611693 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.611711 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.611726 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.715291 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.715374 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.715389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.715415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.715436 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.819019 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.819084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.819095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.819108 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.819118 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.921696 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.921771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.921793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.921822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:23 crc kubenswrapper[5035]: I1002 09:28:23.921844 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:23Z","lastTransitionTime":"2025-10-02T09:28:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.025719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.025772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.025784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.025802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.025816 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.128347 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.128389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.128397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.128411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.128427 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.231066 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.231103 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.231111 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.231125 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.231135 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.335176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.335241 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.335253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.335267 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.335279 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.437603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.437645 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.437657 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.437672 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.437684 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.540434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.540487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.540498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.540520 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.540556 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.643989 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.644061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.644079 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.644104 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.644128 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.752267 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.752358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.752385 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.752503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.752569 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.856073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.856115 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.856126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.856141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.856152 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.958160 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.958193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.958205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.958221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:24 crc kubenswrapper[5035]: I1002 09:28:24.958232 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:24Z","lastTransitionTime":"2025-10-02T09:28:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.061515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.061593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.061619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.061650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.061666 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.162768 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.162867 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.162919 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:25 crc kubenswrapper[5035]: E1002 09:28:25.162975 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:25 crc kubenswrapper[5035]: E1002 09:28:25.163055 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.163120 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:25 crc kubenswrapper[5035]: E1002 09:28:25.163151 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:25 crc kubenswrapper[5035]: E1002 09:28:25.163208 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.164018 5035 scope.go:117] "RemoveContainer" containerID="1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.164394 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.164429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.164439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.164459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.164479 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.267857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.268039 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.268056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.268078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.268094 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.371776 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.371836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.371854 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.371879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.371897 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.474437 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.474486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.474498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.474518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.474571 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.577556 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.577983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.577993 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.578007 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.578016 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.681041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.681121 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.681131 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.681144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.681152 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.783913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.783962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.783973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.783990 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.784003 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.889101 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.889132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.889143 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.889158 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.889170 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.991316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.991355 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.991364 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.991377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:25 crc kubenswrapper[5035]: I1002 09:28:25.991390 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:25Z","lastTransitionTime":"2025-10-02T09:28:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.093717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.093761 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.093781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.093799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.093809 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.177837 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.187258 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.196091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.196128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.196136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.196150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.196160 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.203271 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.218229 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.232872 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.244053 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.260426 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.272946 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.285673 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.298246 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.298426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.298469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.298482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.298497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.298510 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.311463 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.322244 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.344211 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.361336 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.368003 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/2.log" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.368618 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/1.log" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.371562 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0" exitCode=1 Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.371577 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.371692 5035 scope.go:117] "RemoveContainer" containerID="1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.373174 5035 scope.go:117] "RemoveContainer" containerID="d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0" Oct 02 09:28:26 crc kubenswrapper[5035]: E1002 09:28:26.373481 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.377978 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.390812 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.401145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.401209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.401290 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.401323 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.401349 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.401607 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.414280 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.425650 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.444268 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.467397 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.484250 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.494665 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.503387 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.503418 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.503426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.503439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.503448 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.507717 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.519507 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.531194 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.549022 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.564367 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.576394 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.590383 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.603639 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.605626 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.605804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.605814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.606080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.606097 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.621252 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.631207 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.649762 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:26Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.708943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.708985 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.709012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.709024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.709034 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.812201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.812246 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.812255 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.812269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.812279 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.914216 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.914259 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.914271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.914288 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:26 crc kubenswrapper[5035]: I1002 09:28:26.914299 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:26Z","lastTransitionTime":"2025-10-02T09:28:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.016288 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.016326 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.016375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.016391 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.016403 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.103904 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.104022 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:28:59.104000724 +0000 UTC m=+104.460344749 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.104086 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.104215 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.104263 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:59.104252851 +0000 UTC m=+104.460596876 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.119419 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.119456 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.119464 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.119482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.119491 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.162500 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.162617 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.162680 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.162699 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.162559 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.162833 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.162873 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.163021 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.204727 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.204778 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.204799 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.204878 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.204893 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.204921 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:59.204908893 +0000 UTC m=+104.561252918 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.204926 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.204945 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.204989 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.205017 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.205033 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.204997 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:59.204979385 +0000 UTC m=+104.561323430 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:27 crc kubenswrapper[5035]: E1002 09:28:27.205096 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:28:59.205080778 +0000 UTC m=+104.561424853 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.222120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.222163 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.222179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.222200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.222216 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.324893 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.324979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.324991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.325008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.325019 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.375979 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/2.log" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.427297 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.427325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.427337 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.427351 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.427362 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.529225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.529274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.529290 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.529313 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.529329 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.632072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.632119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.632130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.632147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.632160 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.734710 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.734739 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.734747 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.734759 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.734767 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.836920 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.836980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.836993 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.837013 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.837030 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.939142 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.939180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.939190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.939207 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:27 crc kubenswrapper[5035]: I1002 09:28:27.939218 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:27Z","lastTransitionTime":"2025-10-02T09:28:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.041573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.041624 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.041633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.041649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.041658 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.143663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.143728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.143737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.143814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.143935 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.246078 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.246116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.246129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.246145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.246155 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.349352 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.349405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.349418 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.349438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.349453 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.453885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.453914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.453923 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.453936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.453945 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.557444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.557490 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.557506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.557569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.557587 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.659410 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.659469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.659486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.659507 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.659525 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.761636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.761676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.761685 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.761700 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.761713 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.864401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.864436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.864445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.864458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.864467 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.966937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.966969 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.966976 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.966989 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:28 crc kubenswrapper[5035]: I1002 09:28:28.966997 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:28Z","lastTransitionTime":"2025-10-02T09:28:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.070128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.070173 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.070185 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.070202 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.070216 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.161995 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.162089 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.162224 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:29 crc kubenswrapper[5035]: E1002 09:28:29.162108 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:29 crc kubenswrapper[5035]: E1002 09:28:29.162230 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.161996 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:29 crc kubenswrapper[5035]: E1002 09:28:29.162393 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:29 crc kubenswrapper[5035]: E1002 09:28:29.162474 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.172373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.172421 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.172434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.172449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.172460 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.275107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.275168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.275180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.275202 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.275214 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.377917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.377970 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.377979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.377995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.378005 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.480636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.481120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.481152 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.481178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.481206 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.584006 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.584044 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.584056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.584073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.584084 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.686521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.686589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.686601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.686618 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.686628 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.788980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.789015 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.789041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.789053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.789062 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.891297 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.891334 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.891346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.891362 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.891372 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.994416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.994466 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.994475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.994488 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:29 crc kubenswrapper[5035]: I1002 09:28:29.994497 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:29Z","lastTransitionTime":"2025-10-02T09:28:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.098183 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.098228 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.098239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.098256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.098268 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.200590 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.200855 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.200926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.201054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.201136 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.303483 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.303525 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.303562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.303578 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.303589 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.406587 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.406995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.407300 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.407569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.407741 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.510692 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.511000 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.511123 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.511290 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.511416 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.614394 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.614428 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.614438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.614452 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.614460 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.717429 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.717855 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.717885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.717915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.717936 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.820401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.820436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.820446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.820461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.820471 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.922449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.923008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.923085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.923151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:30 crc kubenswrapper[5035]: I1002 09:28:30.923218 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:30Z","lastTransitionTime":"2025-10-02T09:28:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.025671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.025700 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.025709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.025721 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.025729 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.127401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.127434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.127445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.127460 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.127471 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.162185 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.162185 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.162231 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.162277 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.162378 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.162498 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.162586 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.162634 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.229736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.229800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.229818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.229842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.229862 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.332072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.332122 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.332130 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.332148 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.332159 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.434894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.434937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.434946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.434978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.434987 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.538129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.538222 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.538235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.538253 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.538266 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.643966 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.644026 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.644042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.644060 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.644072 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.747235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.747275 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.747288 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.747303 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.747315 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.829119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.829178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.829189 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.829207 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.829222 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.843073 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.846952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.846996 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.847012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.847033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.847047 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.862840 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.867421 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.867471 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.867484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.867502 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.867513 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.883062 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.887027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.887067 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.887084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.887100 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.887110 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.901071 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.904984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.905031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.905050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.905071 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.905088 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.919185 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:31Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:31 crc kubenswrapper[5035]: E1002 09:28:31.919347 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.920944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.920983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.920996 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.921013 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:31 crc kubenswrapper[5035]: I1002 09:28:31.921027 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:31Z","lastTransitionTime":"2025-10-02T09:28:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.023666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.023746 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.023767 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.023795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.023817 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.127060 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.127121 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.127142 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.127166 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.127182 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.230109 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.230160 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.230174 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.230195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.230211 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.333256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.333307 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.333317 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.333331 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.333340 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.435353 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.435406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.435415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.435434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.435452 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.538293 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.538344 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.538356 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.538373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.538385 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.640405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.640453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.640462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.640477 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.640490 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.742709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.742756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.742766 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.742780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.742791 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.844815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.844894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.844908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.844923 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.844934 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.947874 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.947920 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.947932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.947949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:32 crc kubenswrapper[5035]: I1002 09:28:32.947961 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:32Z","lastTransitionTime":"2025-10-02T09:28:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.050245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.050274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.050283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.050296 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.050304 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.152509 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.152568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.152578 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.152592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.152603 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.161967 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.161975 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.161983 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:33 crc kubenswrapper[5035]: E1002 09:28:33.162056 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.162113 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:33 crc kubenswrapper[5035]: E1002 09:28:33.162273 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:33 crc kubenswrapper[5035]: E1002 09:28:33.162297 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:33 crc kubenswrapper[5035]: E1002 09:28:33.162431 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.255451 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.255493 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.255502 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.255518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.255550 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.358186 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.358232 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.358245 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.358261 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.358273 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.460936 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.461010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.461033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.461061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.461088 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.568281 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.568376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.568404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.568436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.568468 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.672244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.672281 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.672292 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.672308 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.672321 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.775086 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.775761 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.775973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.776188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.776578 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.879147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.879470 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.879582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.879675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.879756 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.982153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.982501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.982610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.982687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:33 crc kubenswrapper[5035]: I1002 09:28:33.982749 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:33Z","lastTransitionTime":"2025-10-02T09:28:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.084808 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.084835 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.084843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.084856 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.084865 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.188334 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.188377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.188389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.188405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.188416 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.290695 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.290744 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.290756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.290773 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.290786 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.393090 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.393140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.393151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.393163 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.393171 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.495648 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.495688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.495699 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.495716 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.495729 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.598568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.598625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.598647 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.598667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.598680 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.700617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.700685 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.700702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.700727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.700741 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.803573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.803638 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.803655 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.803684 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.803702 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.905838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.905877 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.905888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.905902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:34 crc kubenswrapper[5035]: I1002 09:28:34.905910 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:34Z","lastTransitionTime":"2025-10-02T09:28:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.008129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.008180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.008188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.008204 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.008215 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.111759 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.111842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.111868 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.111898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.111917 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.162470 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.162559 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.162498 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.162497 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:35 crc kubenswrapper[5035]: E1002 09:28:35.162638 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:35 crc kubenswrapper[5035]: E1002 09:28:35.162699 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:35 crc kubenswrapper[5035]: E1002 09:28:35.162754 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:35 crc kubenswrapper[5035]: E1002 09:28:35.162879 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.213928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.213962 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.213970 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.213983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.213992 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.316197 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.316240 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.316250 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.316265 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.316275 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.418635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.418674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.418690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.418707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.418718 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.521465 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.521506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.521514 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.521548 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.521566 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.624142 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.624180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.624195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.624214 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.624228 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.727620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.728087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.728112 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.728139 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.728159 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.831573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.831643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.831667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.831694 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.831716 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.934263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.934303 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.934311 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.934323 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:35 crc kubenswrapper[5035]: I1002 09:28:35.934332 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:35Z","lastTransitionTime":"2025-10-02T09:28:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.036489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.036525 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.036561 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.036580 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.036593 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.139193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.139234 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.139243 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.139256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.139266 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.205318 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.218607 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.232781 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.241368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.241411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.241426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.241440 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.241450 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.244124 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.256582 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.268207 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.282312 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.295172 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.307499 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.317057 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.334014 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd96bc46c61b0acf0757d7a4e955103e22aeb4ae48811d9c1998c1bac06529f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:12Z\\\",\\\"message\\\":\\\"ics-daemon-hzsjk\\\\\\\", UID:\\\\\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\\\\\", APIVersion:\\\\\\\"v1\\\\\\\", ResourceVersion:\\\\\\\"26881\\\\\\\", FieldPath:\\\\\\\"\\\\\\\"}): type: 'Warning' reason: 'ErrorAddingResource' addLogicalPort failed for openshift-multus/network-metrics-daemon-hzsjk: failed to update pod openshift-multus/network-metrics-daemon-hzsjk: Internal error occurred: failed calling webhook \\\\\\\"pod.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/pod?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:12Z is after 2025-08-24T17:21:41Z\\\\nI1002 09:28:12.534299 6573 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 09:28:12.534383 6573 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 09:28:12.534550 6573 factory.go:656] Stopping watch factory\\\\nI1002 09:28:12.534638 6573 ovnkube.go:599] Stopped ovnkube\\\\nI1002 09:28:12.534437 6573 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 09:28:12.534608 6573 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 09:28:12.534770 6573 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 09:28:12.534943 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.343622 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.344329 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.344365 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.344374 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.344388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.344398 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.355198 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.367466 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.378941 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.390921 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.406867 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:36Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.446625 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.446678 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.446691 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.446709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.446722 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.548631 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.548658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.548667 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.548680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.548689 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.650991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.651042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.651054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.651075 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.651088 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.753240 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.753276 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.753286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.753299 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.753308 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.857232 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.857286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.857300 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.857322 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.857336 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.960756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.960807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.960824 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.960845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:36 crc kubenswrapper[5035]: I1002 09:28:36.960863 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:36Z","lastTransitionTime":"2025-10-02T09:28:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.063894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.063941 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.063959 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.063983 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.064000 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.162811 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:37 crc kubenswrapper[5035]: E1002 09:28:37.163002 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.163024 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.162829 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:37 crc kubenswrapper[5035]: E1002 09:28:37.163160 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:37 crc kubenswrapper[5035]: E1002 09:28:37.163356 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.163681 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:37 crc kubenswrapper[5035]: E1002 09:28:37.163900 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.169325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.169643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.169674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.169704 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.169726 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.273168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.273223 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.273235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.273290 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.273303 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.376222 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.376283 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.376302 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.376326 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.376346 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.480354 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.480432 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.480453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.480482 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.480500 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.583184 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.583344 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.583375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.583446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.583474 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.686168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.686246 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.686264 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.686288 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.686305 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.788707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.788756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.788769 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.788792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.788806 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.892258 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.892329 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.892346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.892372 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.892388 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.994591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.994619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.994644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.994658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:37 crc kubenswrapper[5035]: I1002 09:28:37.994668 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:37Z","lastTransitionTime":"2025-10-02T09:28:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.097737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.097789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.097805 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.097828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.097845 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.201655 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.201757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.201772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.201799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.201817 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.304677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.304731 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.304750 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.304771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.304785 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.407755 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.407797 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.407807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.407824 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.407835 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.510916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.511088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.511107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.511128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.511141 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.614147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.614188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.614196 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.614214 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.614222 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.717956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.718010 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.718031 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.718056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.718071 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.821009 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.821070 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.821087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.821110 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.821128 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.923801 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.923853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.923865 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.923884 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:38 crc kubenswrapper[5035]: I1002 09:28:38.923895 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:38Z","lastTransitionTime":"2025-10-02T09:28:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.027045 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.027093 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.027104 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.027120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.027132 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.131382 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.131463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.131486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.131515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.131574 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.162396 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:39 crc kubenswrapper[5035]: E1002 09:28:39.162680 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.162768 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.162795 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.162414 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:39 crc kubenswrapper[5035]: E1002 09:28:39.163331 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:39 crc kubenswrapper[5035]: E1002 09:28:39.163482 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:39 crc kubenswrapper[5035]: E1002 09:28:39.163582 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.163868 5035 scope.go:117] "RemoveContainer" containerID="d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0" Oct 02 09:28:39 crc kubenswrapper[5035]: E1002 09:28:39.164137 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.180796 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.196106 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.209894 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.220609 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.230953 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.236927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.236980 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.236998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.237021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.237065 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.248077 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.266490 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.282309 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.295790 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.320025 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.335137 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.339065 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.339107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.339119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.339137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.339148 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.352259 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.368661 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.385763 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.402771 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.427312 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.441387 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.441427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.441441 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.441462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.441476 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.443182 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:39Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.543359 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:39 crc kubenswrapper[5035]: E1002 09:28:39.543584 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:39 crc kubenswrapper[5035]: E1002 09:28:39.543664 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:29:11.543640796 +0000 UTC m=+116.899984831 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.543976 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.544023 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.544043 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.544072 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.544097 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.647026 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.647066 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.647079 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.647100 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.647114 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.750077 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.750126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.750141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.750166 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.750227 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.854019 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.854085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.854104 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.854162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.854184 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.956826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.956862 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.956873 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.956889 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:39 crc kubenswrapper[5035]: I1002 09:28:39.956900 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:39Z","lastTransitionTime":"2025-10-02T09:28:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.059764 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.059808 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.059818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.059833 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.059842 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.173736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.173777 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.173789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.173805 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.173817 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.276720 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.276759 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.276789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.276804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.276814 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.379437 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.379470 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.379480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.379494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.379503 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.481994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.482037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.482049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.482068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.482080 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.585323 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.585784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.585992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.586160 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.586308 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.688727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.688768 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.688780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.688795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.688805 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.791634 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.792087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.792302 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.792723 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.793028 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.895880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.896115 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.896237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.896358 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.896708 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:40Z","lastTransitionTime":"2025-10-02T09:28:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.999316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.999689 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:40 crc kubenswrapper[5035]: I1002 09:28:40.999833 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.000024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.000141 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.103004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.103045 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.103054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.103067 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.103075 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.168079 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.168262 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:41 crc kubenswrapper[5035]: E1002 09:28:41.168466 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.168795 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.168907 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:41 crc kubenswrapper[5035]: E1002 09:28:41.169072 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:41 crc kubenswrapper[5035]: E1002 09:28:41.169806 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:41 crc kubenswrapper[5035]: E1002 09:28:41.169987 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.205063 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.205394 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.205476 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.205632 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.205731 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.309074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.309128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.309180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.309206 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.309228 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.412573 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.412638 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.412659 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.412688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.412708 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.515798 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.515885 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.515911 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.515943 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.515965 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.619772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.620164 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.620295 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.620454 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.620665 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.723673 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.723748 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.723764 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.723789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.723805 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.827171 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.827234 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.827246 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.827273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.827287 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.930396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.930459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.930475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.930503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.930519 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.951141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.951186 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.951195 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.951209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.951219 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: E1002 09:28:41.963288 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.966401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.966435 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.966446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.966462 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.966472 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: E1002 09:28:41.977672 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.983004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.983042 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.983053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.983069 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.983080 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:41 crc kubenswrapper[5035]: E1002 09:28:41.994176 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:41Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.997375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.997398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.997407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.997421 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:41 crc kubenswrapper[5035]: I1002 09:28:41.997429 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:41Z","lastTransitionTime":"2025-10-02T09:28:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: E1002 09:28:42.007894 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:42Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.010687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.010717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.010726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.010740 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.010748 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: E1002 09:28:42.020585 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:42Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:42 crc kubenswrapper[5035]: E1002 09:28:42.020850 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.032984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.033022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.033035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.033052 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.033063 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.135956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.136032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.136059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.136089 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.136111 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.239726 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.239795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.239816 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.239843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.239865 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.343594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.343643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.343656 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.343676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.343689 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.446730 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.446796 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.446814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.446848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.446871 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.549574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.549626 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.549642 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.549668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.549685 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.652838 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.652891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.652908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.652932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.652949 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.756002 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.756073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.756083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.756103 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.756119 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.860388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.860463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.860480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.860517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.860556 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.963440 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.963488 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.963505 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.963524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:42 crc kubenswrapper[5035]: I1002 09:28:42.963572 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:42Z","lastTransitionTime":"2025-10-02T09:28:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.075274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.075314 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.075328 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.075349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.075361 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.162895 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.162939 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.163048 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:43 crc kubenswrapper[5035]: E1002 09:28:43.163168 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.163226 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:43 crc kubenswrapper[5035]: E1002 09:28:43.163400 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:43 crc kubenswrapper[5035]: E1002 09:28:43.163816 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:43 crc kubenswrapper[5035]: E1002 09:28:43.164120 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.178701 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.178747 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.178765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.178790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.178811 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.281506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.281574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.281589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.281610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.281624 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.387459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.387508 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.387522 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.387567 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.387584 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.490145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.490176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.490188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.490205 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.490216 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.593053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.593147 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.593204 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.593236 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.593307 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.696771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.696809 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.696817 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.696863 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.696877 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.798693 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.798719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.798728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.798739 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.798747 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.901853 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.901944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.901954 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.901974 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:43 crc kubenswrapper[5035]: I1002 09:28:43.901985 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:43Z","lastTransitionTime":"2025-10-02T09:28:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.004849 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.004898 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.004913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.004933 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.004948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.106890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.106967 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.106986 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.107007 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.107022 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.209736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.209807 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.209824 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.209847 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.209866 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.312909 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.312950 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.312961 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.312977 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.312988 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.415168 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.415212 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.415268 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.415294 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.415310 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.541449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.541495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.541565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.541585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.541596 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.644082 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.644128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.644144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.644166 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.644183 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.746869 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.747000 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.747081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.747118 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.747142 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.849953 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.849991 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.850003 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.850018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.850028 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.952984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.953050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.953162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.953196 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:44 crc kubenswrapper[5035]: I1002 09:28:44.953219 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:44Z","lastTransitionTime":"2025-10-02T09:28:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.056352 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.056405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.056419 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.056438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.056449 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.159791 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.159918 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.159935 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.159953 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.159965 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.162434 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:45 crc kubenswrapper[5035]: E1002 09:28:45.162906 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.162564 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.162517 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:45 crc kubenswrapper[5035]: E1002 09:28:45.163087 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.162592 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:45 crc kubenswrapper[5035]: E1002 09:28:45.163278 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:45 crc kubenswrapper[5035]: E1002 09:28:45.163418 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.174082 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.262567 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.262666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.262675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.262688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.262700 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.367569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.367628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.367648 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.367674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.367691 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.471030 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.471080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.471094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.471111 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.471122 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.575946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.575987 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.576004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.576024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.576039 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.678978 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.679037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.679057 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.679079 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.679097 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.782756 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.782814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.782827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.782843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.782855 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.885416 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.885453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.885464 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.885481 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.885492 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.989212 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.989276 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.989296 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.989324 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:45 crc kubenswrapper[5035]: I1002 09:28:45.989345 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:45Z","lastTransitionTime":"2025-10-02T09:28:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.093074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.093172 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.093193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.093221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.093250 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.175974 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.186911 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.209135 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.209167 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.209177 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.209193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.209215 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.211843 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.234689 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.247927 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.257345 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.271959 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.283012 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.293633 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.306168 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.311094 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.311136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.311144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.311156 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.311168 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.320046 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.331126 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.340983 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.359118 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.370661 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.384964 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.398578 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.411609 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:46Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.413518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.413570 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.413582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.413598 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.413610 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.518200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.518266 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.518290 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.518324 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.518348 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.622122 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.622188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.622200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.622219 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.622236 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.725574 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.725645 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.725664 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.725688 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.725706 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.830695 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.830762 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.830779 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.830806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.830839 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.934360 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.934458 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.934478 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.934502 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:46 crc kubenswrapper[5035]: I1002 09:28:46.934517 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:46Z","lastTransitionTime":"2025-10-02T09:28:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.037095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.037144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.037155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.037175 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.037186 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.143487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.143550 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.143562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.143579 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.143589 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.162151 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.162206 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:47 crc kubenswrapper[5035]: E1002 09:28:47.162288 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.162372 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:47 crc kubenswrapper[5035]: E1002 09:28:47.162398 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.162504 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:47 crc kubenswrapper[5035]: E1002 09:28:47.162616 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:47 crc kubenswrapper[5035]: E1002 09:28:47.162813 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.246982 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.247020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.247033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.247050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.247065 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.350174 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.350251 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.350269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.350298 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.350326 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.440842 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/0.log" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.440929 5035 generic.go:334] "Generic (PLEG): container finished" podID="5fafe5e6-83b2-447b-9379-b26e8071166b" containerID="d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e" exitCode=1 Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.440976 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerDied","Data":"d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.441668 5035 scope.go:117] "RemoveContainer" containerID="d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.454745 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.454772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.454780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.454794 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.454804 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.461614 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.482049 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:46Z\\\",\\\"message\\\":\\\"2025-10-02T09:28:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6\\\\n2025-10-02T09:28:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6 to /host/opt/cni/bin/\\\\n2025-10-02T09:28:01Z [verbose] multus-daemon started\\\\n2025-10-02T09:28:01Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:28:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.492835 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.517939 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.531486 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.552886 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.557994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.558041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.558053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.558074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.558086 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.570356 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.587996 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.604458 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.621712 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.638586 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.657669 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.661456 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.661495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.661517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.661562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.661578 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.675265 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.692514 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.705485 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.719277 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.733099 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.746116 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:47Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.764138 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.764180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.764189 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.764201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.764210 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.866110 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.866272 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.866391 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.866487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.866596 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.972610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.972649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.972663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.972676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:47 crc kubenswrapper[5035]: I1002 09:28:47.972685 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:47Z","lastTransitionTime":"2025-10-02T09:28:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.075729 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.075772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.075789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.075812 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.075828 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.178693 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.178740 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.178758 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.178783 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.178801 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.281097 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.281137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.281148 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.281163 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.281174 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.383475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.383518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.383543 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.383557 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.383567 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.447596 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/0.log" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.447697 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerStarted","Data":"bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.466337 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.479597 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:46Z\\\",\\\"message\\\":\\\"2025-10-02T09:28:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6\\\\n2025-10-02T09:28:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6 to /host/opt/cni/bin/\\\\n2025-10-02T09:28:01Z [verbose] multus-daemon started\\\\n2025-10-02T09:28:01Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:28:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.486440 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.486560 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.486577 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.486602 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.486614 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.492944 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.515041 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.526835 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.542081 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.556330 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.581927 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.589102 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.589166 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.589180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.589199 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.589211 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.599255 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.616373 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.631053 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.645638 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.656491 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.669667 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.684910 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.690981 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.691012 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.691025 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.691041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.691052 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.697425 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.713689 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.729406 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:48Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.794299 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.794381 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.794404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.794436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.794456 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.897877 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.898318 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.898431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.898615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:48 crc kubenswrapper[5035]: I1002 09:28:48.898780 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:48Z","lastTransitionTime":"2025-10-02T09:28:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.001712 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.002033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.002126 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.002225 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.002321 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.105415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.105495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.105509 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.105527 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.105561 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.162319 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.162336 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:49 crc kubenswrapper[5035]: E1002 09:28:49.162566 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.162336 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:49 crc kubenswrapper[5035]: E1002 09:28:49.162688 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.162344 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:49 crc kubenswrapper[5035]: E1002 09:28:49.162449 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:49 crc kubenswrapper[5035]: E1002 09:28:49.162753 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.208282 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.208344 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.208356 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.208370 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.208380 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.310307 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.310647 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.310722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.310897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.310964 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.413505 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.413565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.413577 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.413593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.413604 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.516155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.516208 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.516220 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.516239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.516250 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.619087 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.619145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.619155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.619170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.619179 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.721563 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.721594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.721604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.721619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.721628 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.823997 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.824047 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.824059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.824074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.824086 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.926568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.926623 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.926639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.926661 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:49 crc kubenswrapper[5035]: I1002 09:28:49.926676 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:49Z","lastTransitionTime":"2025-10-02T09:28:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.029973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.030101 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.030120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.030145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.030162 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.132731 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.132791 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.132806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.132829 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.132846 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.235340 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.235409 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.235419 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.235434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.235445 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.338737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.338781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.338789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.338802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.338810 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.441176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.441222 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.441250 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.441265 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.441275 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.544736 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.544790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.544801 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.544817 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.544827 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.647440 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.647503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.647517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.647565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.647663 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.750214 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.750278 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.750296 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.750320 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.750337 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.854324 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.854375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.854389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.854406 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.854420 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.957473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.957582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.957596 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.957613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:50 crc kubenswrapper[5035]: I1002 09:28:50.957624 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:50Z","lastTransitionTime":"2025-10-02T09:28:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.059930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.060182 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.060251 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.060320 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.060382 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.161962 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:51 crc kubenswrapper[5035]: E1002 09:28:51.162091 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.162112 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.162125 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:51 crc kubenswrapper[5035]: E1002 09:28:51.162183 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:51 crc kubenswrapper[5035]: E1002 09:28:51.162278 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.162653 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:51 crc kubenswrapper[5035]: E1002 09:28:51.162725 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.163169 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.163201 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.163219 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.163238 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.163253 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.266342 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.266404 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.266423 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.266449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.266469 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.370193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.370237 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.370246 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.370260 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.370269 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.472782 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.473106 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.473211 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.473307 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.473392 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.575805 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.575848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.575860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.575877 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.575892 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.678015 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.678058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.678074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.678091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.678100 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.780459 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.780508 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.780524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.780583 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.780598 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.886851 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.887177 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.887267 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.887375 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.887481 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.990135 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.990184 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.990200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.990221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:51 crc kubenswrapper[5035]: I1002 09:28:51.990236 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:51Z","lastTransitionTime":"2025-10-02T09:28:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.056033 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.056080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.056095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.056112 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.056121 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: E1002 09:28:52.071477 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:52Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.076447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.076479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.076488 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.076501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.076510 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: E1002 09:28:52.087461 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:52Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.091749 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.092032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.092244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.092489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.092689 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: E1002 09:28:52.104898 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:52Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.109256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.109298 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.109307 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.109322 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.109331 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: E1002 09:28:52.119346 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:52Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.122903 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.122952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.122969 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.122992 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.123009 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: E1002 09:28:52.136088 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:52Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:52 crc kubenswrapper[5035]: E1002 09:28:52.136314 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.138415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.138455 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.138469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.138484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.138493 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.177725 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.241346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.241382 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.241393 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.241409 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.241422 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.343517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.343601 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.343619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.343646 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.343666 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.446702 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.446815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.446850 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.446879 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.446899 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.549952 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.550009 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.550028 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.550053 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.550071 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.653631 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.653703 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.653727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.653755 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.653779 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.757038 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.757085 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.757098 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.757113 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.757124 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.858866 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.858908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.858919 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.858935 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.858947 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.961278 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.961313 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.961325 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.961341 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:52 crc kubenswrapper[5035]: I1002 09:28:52.961353 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:52Z","lastTransitionTime":"2025-10-02T09:28:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.063959 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.064020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.064037 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.064058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.064072 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.162256 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.162256 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:53 crc kubenswrapper[5035]: E1002 09:28:53.162386 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.162489 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:53 crc kubenswrapper[5035]: E1002 09:28:53.162569 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:53 crc kubenswrapper[5035]: E1002 09:28:53.162710 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.163250 5035 scope.go:117] "RemoveContainer" containerID="d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.168871 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.168922 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.168935 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.168948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.168957 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.171057 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:53 crc kubenswrapper[5035]: E1002 09:28:53.171139 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.272179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.272236 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.272257 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.272286 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.272308 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.374582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.374658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.374669 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.374683 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.374694 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.477822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.477858 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.477867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.477881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.477891 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.580489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.580524 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.580549 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.580562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.580571 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.683789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.683851 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.683869 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.683894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.683914 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.787211 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.787273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.787289 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.787312 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.787327 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.890050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.890111 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.890119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.890131 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.890140 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.992677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.992753 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.992777 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.992808 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:53 crc kubenswrapper[5035]: I1002 09:28:53.992832 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:53Z","lastTransitionTime":"2025-10-02T09:28:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.095480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.095592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.095662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.095690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.095712 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.198985 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.199034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.199097 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.199122 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.199135 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.301880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.301922 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.301934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.301950 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.301961 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.404553 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.404592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.404603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.404618 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.404630 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.508627 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.508656 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.508665 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.508679 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.508687 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.611914 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.611994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.612017 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.612044 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.612065 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.714677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.714745 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.714770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.714804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.714827 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.817880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.817915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.817924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.817939 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.817948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.921577 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.921628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.921644 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.921664 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:54 crc kubenswrapper[5035]: I1002 09:28:54.921682 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:54Z","lastTransitionTime":"2025-10-02T09:28:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.024809 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.024857 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.024877 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.024903 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.024918 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.127097 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.127133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.127142 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.127157 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.127168 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.162773 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.162887 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:55 crc kubenswrapper[5035]: E1002 09:28:55.162914 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.162959 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.162962 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:55 crc kubenswrapper[5035]: E1002 09:28:55.163059 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:55 crc kubenswrapper[5035]: E1002 09:28:55.163251 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:55 crc kubenswrapper[5035]: E1002 09:28:55.163303 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.230329 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.230368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.230381 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.230398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.230410 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.332749 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.332791 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.332800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.332814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.332824 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.435591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.435666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.435676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.435694 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.435707 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.474262 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/2.log" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.477102 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.478137 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.500124 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.512144 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:46Z\\\",\\\"message\\\":\\\"2025-10-02T09:28:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6\\\\n2025-10-02T09:28:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6 to /host/opt/cni/bin/\\\\n2025-10-02T09:28:01Z [verbose] multus-daemon started\\\\n2025-10-02T09:28:01Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:28:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.526995 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.538278 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.538341 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.538354 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.538369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.538380 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.545655 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.554811 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.568704 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.582070 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.595896 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.607069 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.618961 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.637556 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6bfae56-821c-49d6-ad3c-6fb9b4b0cf6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd12dcbe766e84270752eb65d14f3e5881640d0e46d94c6c2d695edc83b6915e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f483770dacbda745acb21802f775090ff51df5bc932518311b645c8539c2d5f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd669b07e45dd36d35fbe716218112dd79d06ef543b47b00108be49cfa39efa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6e2066bf1530e2ef3f455e83a2822b8db1d1eaccdb3c199d4748f975771957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5811facbcc46f163c5dc71d7fc6095a509746b50764410ed99c94ebb24c317ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.640333 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.640364 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.640373 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.640388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.640397 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.653409 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.680435 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.690028 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.701997 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.713585 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.724557 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.736793 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.742193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.742224 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.742232 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.742246 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.742257 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.747561 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:55Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.844711 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.844793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.844804 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.844833 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.844849 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.947780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.947813 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.947827 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.947850 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:55 crc kubenswrapper[5035]: I1002 09:28:55.947861 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:55Z","lastTransitionTime":"2025-10-02T09:28:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.051106 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.051169 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.051181 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.051198 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.051211 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.153908 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.153985 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.153998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.154014 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.154028 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.177268 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:46Z\\\",\\\"message\\\":\\\"2025-10-02T09:28:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6\\\\n2025-10-02T09:28:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6 to /host/opt/cni/bin/\\\\n2025-10-02T09:28:01Z [verbose] multus-daemon started\\\\n2025-10-02T09:28:01Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:28:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.187552 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.205233 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.217068 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.230768 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.242591 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.255719 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.256426 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.256452 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.256501 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.256517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.256552 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.270609 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.290890 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.302056 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.324241 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.340356 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.351854 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.362140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.362189 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.362200 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.362221 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.362235 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.375254 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6bfae56-821c-49d6-ad3c-6fb9b4b0cf6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd12dcbe766e84270752eb65d14f3e5881640d0e46d94c6c2d695edc83b6915e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f483770dacbda745acb21802f775090ff51df5bc932518311b645c8539c2d5f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd669b07e45dd36d35fbe716218112dd79d06ef543b47b00108be49cfa39efa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6e2066bf1530e2ef3f455e83a2822b8db1d1eaccdb3c199d4748f975771957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5811facbcc46f163c5dc71d7fc6095a509746b50764410ed99c94ebb24c317ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.388977 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.400281 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.409628 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.422138 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.452618 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.465907 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.465946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.465958 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.465975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.465987 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.482169 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/3.log" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.483328 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/2.log" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.486241 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" exitCode=1 Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.486300 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.486347 5035 scope.go:117] "RemoveContainer" containerID="d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.487126 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:28:56 crc kubenswrapper[5035]: E1002 09:28:56.487385 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.500472 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.514809 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.528843 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.539829 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.551041 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.569118 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.569176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.569191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.569207 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.569218 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.569250 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d3cc8611e88d8c22c4d353f28d145a9ee84a8baaf1d7b704a3a5d165d1b846e0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:26Z\\\",\\\"message\\\":\\\"dminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 09:28:26.203468 6847 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203609 6847 factory.go:656] Stopping watch factory\\\\nI1002 09:28:26.203740 6847 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203785 6847 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.203904 6847 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 09:28:26.203927 6847 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204102 6847 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 09:28:26.204216 6847 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:55Z\\\",\\\"message\\\":\\\"\\\\\\\"TCP\\\\\\\", inport:9979, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:28:55.883897 7176 services_controller.go:444] Built service openshift-etcd/etcd LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:28:55.883903 7176 services_controller.go:445] Built service openshift-etcd/etcd LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:28:55.883837 7176 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 09:28:55.883158 7176 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.579309 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.590138 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.600768 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.611276 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:46Z\\\",\\\"message\\\":\\\"2025-10-02T09:28:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6\\\\n2025-10-02T09:28:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6 to /host/opt/cni/bin/\\\\n2025-10-02T09:28:01Z [verbose] multus-daemon started\\\\n2025-10-02T09:28:01Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:28:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.619346 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.630311 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.641726 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.655354 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.667569 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.671292 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.671332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.671346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.671363 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.671373 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.679740 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.699642 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6bfae56-821c-49d6-ad3c-6fb9b4b0cf6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd12dcbe766e84270752eb65d14f3e5881640d0e46d94c6c2d695edc83b6915e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f483770dacbda745acb21802f775090ff51df5bc932518311b645c8539c2d5f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd669b07e45dd36d35fbe716218112dd79d06ef543b47b00108be49cfa39efa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6e2066bf1530e2ef3f455e83a2822b8db1d1eaccdb3c199d4748f975771957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5811facbcc46f163c5dc71d7fc6095a509746b50764410ed99c94ebb24c317ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.713781 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.723500 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:56Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.773405 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.773438 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.773446 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.773457 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.773466 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.875505 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.875566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.875577 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.875593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.875606 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.978745 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.978796 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.978808 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.978826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:56 crc kubenswrapper[5035]: I1002 09:28:56.978839 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:56Z","lastTransitionTime":"2025-10-02T09:28:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.081075 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.081116 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.081129 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.081145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.081157 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.162303 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.162373 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:57 crc kubenswrapper[5035]: E1002 09:28:57.162481 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.162525 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:57 crc kubenswrapper[5035]: E1002 09:28:57.162768 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.162803 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:57 crc kubenswrapper[5035]: E1002 09:28:57.162842 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:57 crc kubenswrapper[5035]: E1002 09:28:57.163372 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.183327 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.183383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.183400 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.183423 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.183441 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.286613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.286660 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.286676 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.286698 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.286716 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.389709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.389793 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.389819 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.389849 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.389869 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.493008 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.493105 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.493124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.493153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.493174 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.494011 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/3.log" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.498895 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:28:57 crc kubenswrapper[5035]: E1002 09:28:57.499168 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.517145 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:46Z\\\",\\\"message\\\":\\\"2025-10-02T09:28:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6\\\\n2025-10-02T09:28:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6 to /host/opt/cni/bin/\\\\n2025-10-02T09:28:01Z [verbose] multus-daemon started\\\\n2025-10-02T09:28:01Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:28:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.529720 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.563284 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:55Z\\\",\\\"message\\\":\\\"\\\\\\\"TCP\\\\\\\", inport:9979, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:28:55.883897 7176 services_controller.go:444] Built service openshift-etcd/etcd LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:28:55.883903 7176 services_controller.go:445] Built service openshift-etcd/etcd LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:28:55.883837 7176 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 09:28:55.883158 7176 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.575499 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.595010 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.595728 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.595787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.595805 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.595826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.595841 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.606063 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.618411 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.633428 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.645557 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.658152 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.675272 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.687628 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.695965 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.698209 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.698252 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.698271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.698285 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.698296 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.712565 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6bfae56-821c-49d6-ad3c-6fb9b4b0cf6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd12dcbe766e84270752eb65d14f3e5881640d0e46d94c6c2d695edc83b6915e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f483770dacbda745acb21802f775090ff51df5bc932518311b645c8539c2d5f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd669b07e45dd36d35fbe716218112dd79d06ef543b47b00108be49cfa39efa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6e2066bf1530e2ef3f455e83a2822b8db1d1eaccdb3c199d4748f975771957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5811facbcc46f163c5dc71d7fc6095a509746b50764410ed99c94ebb24c317ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.724028 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.733972 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.744617 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.756931 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.770710 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:28:57Z is after 2025-08-24T17:21:41Z" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.800774 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.800820 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.800828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.800846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.800858 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.903481 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.903604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.903623 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.903639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:57 crc kubenswrapper[5035]: I1002 09:28:57.903651 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:57Z","lastTransitionTime":"2025-10-02T09:28:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.005487 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.005520 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.005528 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.005557 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.005576 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.108212 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.108248 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.108258 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.108273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.108283 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.210851 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.210900 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.210913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.210931 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.210965 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.313151 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.313190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.313210 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.313227 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.313239 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.415617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.415660 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.415685 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.415707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.415722 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.517919 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.517969 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.517982 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.517998 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.518010 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.621449 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.621518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.621551 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.621567 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.621577 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.725287 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.725349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.725377 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.725400 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.725415 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.828479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.828558 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.828575 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.828600 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.828615 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.932035 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.932108 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.932124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.932145 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:58 crc kubenswrapper[5035]: I1002 09:28:58.932159 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:58Z","lastTransitionTime":"2025-10-02T09:28:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.034781 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.034823 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.034831 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.034846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.034856 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.138393 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.138433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.138474 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.138495 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.138506 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.162217 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.162257 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.162351 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.162389 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.162241 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.162505 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.162602 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.162729 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.169750 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:03.169715884 +0000 UTC m=+168.526059929 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.169904 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.170081 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.170276 5035 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.170328 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:03.170317562 +0000 UTC m=+168.526661587 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.241356 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.241401 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.241414 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.241430 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.241443 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.271297 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.271359 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.271380 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271483 5035 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271568 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:03.27152378 +0000 UTC m=+168.627867805 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271656 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271667 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271742 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271690 5035 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271765 5035 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271768 5035 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271840 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:03.271822149 +0000 UTC m=+168.628166244 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:59 crc kubenswrapper[5035]: E1002 09:28:59.271860 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:03.27185309 +0000 UTC m=+168.628197115 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.344902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.345188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.345199 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.345214 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.345224 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.449017 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.449089 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.449115 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.449144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.449165 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.552727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.552771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.552784 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.552799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.552811 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.655494 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.655620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.655637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.655662 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.655672 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.758299 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.758343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.758354 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.758368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.758378 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.860979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.861027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.861038 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.861054 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.861066 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.963739 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.963770 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.963780 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.963795 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:28:59 crc kubenswrapper[5035]: I1002 09:28:59.963806 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:28:59Z","lastTransitionTime":"2025-10-02T09:28:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.067339 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.067396 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.067407 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.067432 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.067455 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.169485 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.169554 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.169568 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.169619 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.169633 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.272842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.272915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.272925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.272938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.272948 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.375842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.375896 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.375904 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.375918 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.375928 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.478133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.478170 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.478178 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.478190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.478202 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.580649 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.580683 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.580692 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.580704 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.580712 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.683591 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.683684 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.683707 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.683738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.683760 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.786367 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.786410 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.786419 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.786431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.786441 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.888594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.888635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.888658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.888673 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.888686 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.991427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.991484 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.991502 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.991525 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:00 crc kubenswrapper[5035]: I1002 09:29:00.991576 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:00Z","lastTransitionTime":"2025-10-02T09:29:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.094343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.094408 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.094424 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.094448 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.094465 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.162554 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.162635 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:01 crc kubenswrapper[5035]: E1002 09:29:01.162753 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:01 crc kubenswrapper[5035]: E1002 09:29:01.162827 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.162905 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:01 crc kubenswrapper[5035]: E1002 09:29:01.162988 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.163019 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:01 crc kubenswrapper[5035]: E1002 09:29:01.163077 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.196620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.196663 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.196671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.196685 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.196698 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.299913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.299973 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.299984 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.300005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.300017 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.402924 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.403191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.403267 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.403400 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.403487 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.506783 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.506926 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.506940 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.506960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.506978 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.614081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.614137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.614160 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.614180 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.614193 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.717119 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.717174 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.717190 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.717213 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.717228 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.820757 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.820832 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.820856 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.820889 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.820913 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.922730 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.923024 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.923086 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.923149 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:01 crc kubenswrapper[5035]: I1002 09:29:01.923203 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:01Z","lastTransitionTime":"2025-10-02T09:29:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.025963 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.026683 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.026826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.026933 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.027025 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.130569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.130608 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.130620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.130636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.130648 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.211058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.211115 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.211123 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.211136 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.211145 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: E1002 09:29:02.226643 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.231137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.231191 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.231211 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.231235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.231251 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: E1002 09:29:02.246564 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.251274 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.251308 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.251319 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.251334 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.251344 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: E1002 09:29:02.267836 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.271278 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.271301 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.271308 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.271322 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.271329 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: E1002 09:29:02.281791 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.284760 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.284812 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.284823 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.284837 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.284847 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: E1002 09:29:02.296103 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:02Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:02 crc kubenswrapper[5035]: E1002 09:29:02.296212 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.297453 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.297497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.297513 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.297552 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.297566 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.400276 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.400323 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.400332 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.400346 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.400355 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.511061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.511097 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.511107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.511120 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.511129 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.614772 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.614834 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.614848 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.614864 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.614873 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.717628 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.717683 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.717700 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.717722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.717740 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.820159 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.820203 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.820218 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.820231 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.820239 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.922836 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.922883 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.922899 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.922921 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:02 crc kubenswrapper[5035]: I1002 09:29:02.922937 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:02Z","lastTransitionTime":"2025-10-02T09:29:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.026374 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.026433 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.026463 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.026486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.026502 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.129300 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.129361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.129371 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.129386 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.129398 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.162182 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.162279 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.162279 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.162276 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:03 crc kubenswrapper[5035]: E1002 09:29:03.162413 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:03 crc kubenswrapper[5035]: E1002 09:29:03.162595 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:03 crc kubenswrapper[5035]: E1002 09:29:03.162685 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:03 crc kubenswrapper[5035]: E1002 09:29:03.162891 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.232916 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.232979 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.232995 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.233018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.233036 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.336264 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.336316 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.336327 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.336342 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.336353 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.439580 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.439659 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.439684 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.439719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.439746 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.543108 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.543159 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.543172 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.543192 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.543204 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.645629 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.645665 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.645675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.645690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.645701 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.749133 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.749176 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.749188 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.749202 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.749214 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.852255 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.852582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.852592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.852606 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.852614 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.955299 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.955368 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.955390 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.955423 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:03 crc kubenswrapper[5035]: I1002 09:29:03.955447 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:03Z","lastTransitionTime":"2025-10-02T09:29:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.058860 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.058905 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.058917 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.058930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.058939 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.161513 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.161575 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.161585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.161600 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.161613 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.264303 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.264352 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.264365 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.264384 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.264396 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.367082 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.367186 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.367214 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.367247 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.367270 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.470415 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.470481 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.470497 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.470519 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.470562 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.574109 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.574173 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.574197 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.574226 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.574248 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.677150 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.677219 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.677241 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.677271 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.677293 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.780915 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.780996 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.781020 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.781056 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.781081 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.884402 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.884469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.884486 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.884519 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.884559 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.987505 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.987587 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.987599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.987620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:04 crc kubenswrapper[5035]: I1002 09:29:04.987636 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:04Z","lastTransitionTime":"2025-10-02T09:29:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.090999 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.091058 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.091080 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.091107 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.091127 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.162932 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.163012 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.163084 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.163148 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:05 crc kubenswrapper[5035]: E1002 09:29:05.163163 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:05 crc kubenswrapper[5035]: E1002 09:29:05.163295 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:05 crc kubenswrapper[5035]: E1002 09:29:05.163322 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:05 crc kubenswrapper[5035]: E1002 09:29:05.163374 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.198090 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.198160 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.198174 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.198199 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.198216 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.300894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.300928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.300937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.300950 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.300958 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.403580 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.403623 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.403634 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.403650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.403661 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.505884 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.505927 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.505937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.505951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.505962 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.608189 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.608222 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.608231 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.608247 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.608258 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.710480 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.710549 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.710566 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.710585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.710600 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.812822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.812863 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.812876 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.812894 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.812907 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.916488 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.916593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.916616 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.916642 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:05 crc kubenswrapper[5035]: I1002 09:29:05.916663 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:05Z","lastTransitionTime":"2025-10-02T09:29:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.018634 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.018668 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.018675 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.018687 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.018919 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.122298 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.122356 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.122376 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.122403 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.122423 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.177957 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-j6gm8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"03885de7-3b56-436e-809f-f1ce7c027965\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e765c8bb79ca1948b2ea54f72a2e75af2edb0d43e10d1d3c1d0e0da16a7f886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qdsv6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-j6gm8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.205428 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffeef997-af36-4315-8a12-46ecf69976e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:55Z\\\",\\\"message\\\":\\\"\\\\\\\"TCP\\\\\\\", inport:9979, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1002 09:28:55.883897 7176 services_controller.go:444] Built service openshift-etcd/etcd LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:28:55.883903 7176 services_controller.go:445] Built service openshift-etcd/etcd LB template configs for network=default: []services.lbConfig(nil)\\\\nI1002 09:28:55.883837 7176 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns-operator/metrics]} name:Service_openshift-dns-operator/metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.174:9393:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d937b3b3-82c3-4791-9a66-41b9fed53e9d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1002 09:28:55.883158 7176 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tlfx7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-cmx5n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.221389 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-8xzpk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8222638-7aa2-4722-8013-a87cf909feae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59359f5c94a1e744400c658195e23e1dcddc2e3e1ae845f585b72f42bafa45e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5csb5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:59Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-8xzpk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.225292 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.225337 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.225352 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.225393 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.225408 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.235281 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2b03b9-7b34-478f-ac90-7ef709d28b7b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://857fe1171324034f7ee3058a4e895716dfa2dbe204e0af7fa60544aaef66b722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a855c84a72e7f2eab165f4d2536b382f5b72767921245f225b8e8a696e79948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://284d3d47915ff032c80629335fc5257a1ef6f29614787f05a15253e7fd89b7a3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd3fbfecc4dad9d3fd6fef345fb16a2c966d0cb9ae344074c0414315a5271658\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.253305 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.272086 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8rpfb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fafe5e6-83b2-447b-9379-b26e8071166b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T09:28:46Z\\\",\\\"message\\\":\\\"2025-10-02T09:28:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6\\\\n2025-10-02T09:28:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d63dbac4-acc9-48e1-b4c3-ea228b5682e6 to /host/opt/cni/bin/\\\\n2025-10-02T09:28:01Z [verbose] multus-daemon started\\\\n2025-10-02T09:28:01Z [verbose] Readiness Indicator file check\\\\n2025-10-02T09:28:46Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qfvf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8rpfb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.287801 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.306274 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"198f377e-1ff7-4845-9330-9a7eb67568b9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7825df4ac860b78e7f3842424e2456d03cbe1ac5f4b2c654cad81c9d31ea6a63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f39ff0e3a0d829b1b437e622098638aae07c0cc852c1d20dd59b42b5f2c1c9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb76a47a71028b9a0c0c8f5b9c3e7f498de43c60aeda4c9b8bd44845f61f7cd5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc7e7e77fedfb8d79d3dfa097bc53a12d89a3667a18960151926f202b4676bc5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1aaf0826563047ce21b1e7090dc9b26344eeaf8ecc5d1288345532b697589d6b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T09:27:54Z\\\",\\\"message\\\":\\\"le observer\\\\nW1002 09:27:53.573708 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 09:27:53.573877 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 09:27:53.574796 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3817241443/tls.crt::/tmp/serving-cert-3817241443/tls.key\\\\\\\"\\\\nI1002 09:27:53.903195 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 09:27:53.926309 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 09:27:53.926374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 09:27:53.926404 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 09:27:53.926411 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 09:27:53.937950 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1002 09:27:53.937988 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 09:27:53.938013 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 09:27:53.938030 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 09:27:53.938039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 09:27:53.938043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 09:27:53.938046 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 09:27:53.940606 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed693a1afcba8b6db9f2d663d39da50ee8c743ef9b2d0167cc3e8748b19be81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d512613401ddd19b19b41ff9d26f330e3c265030eca99c40395e5017962c4dcc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.327144 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"67c83d64-e4d4-44b0-81a0-900c2a83ed0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4203abda040b2e95947f964f3a53aed4e9166de9e2ef2f7ccaec4d033698a2f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aed5030f2aaae40e8bb02220019e2b7a805136446bed38f8396352f3f4dfade5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://af708c34b78007660dbd2485c96949d00374e584804a4e5ebec78d883c6eb681\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7700258d5db67a39bbd4f83e9d4c16e81e75c02514e888c10a081ede27d7e783\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.328938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.329367 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.329423 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.329448 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.329461 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.349738 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0aa458f9332d8924a59a64d96646d92070fd4076e98789b64e8febe950e08e7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.365604 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"452af00b-602d-43ab-a345-5453d6aebcf0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c2vpf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hzsjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.381305 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a04e9ab3-88f7-45a9-9be7-cdd5f9fc1aff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e388ade8267cb9cd4815883d3aaf01440621cbe506a0921280ca1983b685890d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a53022f100589d38de8a34e0948c0c1112ba7850a124d27889bfdda39a5f5927\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.407290 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6bfae56-821c-49d6-ad3c-6fb9b4b0cf6c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd12dcbe766e84270752eb65d14f3e5881640d0e46d94c6c2d695edc83b6915e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f483770dacbda745acb21802f775090ff51df5bc932518311b645c8539c2d5f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3dd669b07e45dd36d35fbe716218112dd79d06ef543b47b00108be49cfa39efa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6e2066bf1530e2ef3f455e83a2822b8db1d1eaccdb3c199d4748f975771957\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5811facbcc46f163c5dc71d7fc6095a509746b50764410ed99c94ebb24c317ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c62f3af9cbdf0c63ea989907e17393330fb3ec99b75bcef2d2ab8f7a4624ee18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb070743c8030ad362e927451c33b900291f20fdef97cebb2ed44555737d69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:32Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70058e394651aea48e5fd856ccad539586ace605998e4caa24abfd2f62321205\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:16Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.426217 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-4hwst" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"271a556b-657c-4069-8cbc-a091395ca2ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dadf507ce6d4985f643f7eabb8b4769daff77dd64285590b55f2a7707c6d224f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d998327165d6761c7cf7f53948dcbac59fbcefcedf17be39c5a19857aef71a32\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:27:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://95a3c7549294885ed9dc56237255c6e0dce739b0443898c407d8e3f87e76aa8c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:27:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d040fd09cd634846f6a4a863d8a56726f446847fb01921c477edd656dd4de925\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fff7936b990cc3d971cf5d7297c3f47aafef9b4c06219c31c91aecebe08cc7e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a6d7bfb6ae787eb9d4cffa193abe19ea277b16eb135914862d73a8246de69bb8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a6cac42b261abfb65e126dfecc2a7ed6b99f1613e23b94cdf939afbebaa79d9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T09:28:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fcbgk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-4hwst\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.431609 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.431645 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.431655 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.431671 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.431683 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.439335 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f04b76-b225-449c-adc6-e979f26ab0da\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe038d2fc24420c6a2871210abddb0ea4e94f59b96e860a576743ee00848f81d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zfjd9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:27:55Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-k6b5v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.453913 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88a34b1e-c1fe-435c-9cfa-3288d778139c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T09:28:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf46a2bbfd77e71f2e98642d4e8aa2eb0018f5713592fdcc1402db7b48a9889d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7069df8f58234599f4b159b8cd33faf33fed10646d222f7f1db6a9e0201520d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:28:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7kxwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T09:28:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-5k5w7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.469858 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.485851 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ea047e015f0eec806a68cdffd691de16b9911ece988b892c178d15fbbe0ce5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff8fa9966da5d6f8ee966d90ab093748e0e8156ca5099c67b38dd8010dceabde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.499073 5035 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T09:27:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7109ad0142572778dbe75192295a1a9d7ae37fa0864bfc27af1e7c6790ca10fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T09:27:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:06Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.533867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.533899 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.533909 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.533921 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.533930 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.636717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.636775 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.636792 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.636815 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.636833 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.739951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.740027 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.740068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.740102 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.740125 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.843244 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.843306 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.843315 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.843330 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.843341 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.945882 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.945932 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.945949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.945971 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:06 crc kubenswrapper[5035]: I1002 09:29:06.945988 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:06Z","lastTransitionTime":"2025-10-02T09:29:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.048970 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.049034 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.049049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.049074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.049090 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.151893 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.151929 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.151938 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.151951 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.151960 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.162368 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.162488 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:07 crc kubenswrapper[5035]: E1002 09:29:07.162584 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:07 crc kubenswrapper[5035]: E1002 09:29:07.162762 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.162884 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:07 crc kubenswrapper[5035]: E1002 09:29:07.162977 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.163045 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:07 crc kubenswrapper[5035]: E1002 09:29:07.163115 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.255164 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.255216 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.255234 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.255254 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.255265 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.357471 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.357565 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.357582 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.357605 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.357620 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.460351 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.460390 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.460398 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.460411 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.460420 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.562822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.562867 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.562880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.562895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.562906 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.665594 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.665642 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.665653 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.665709 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.665723 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.768073 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.768127 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.768140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.768156 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.768167 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.870765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.870806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.870822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.870842 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.870858 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.973734 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.973785 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.973796 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.973812 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:07 crc kubenswrapper[5035]: I1002 09:29:07.973822 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:07Z","lastTransitionTime":"2025-10-02T09:29:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.076084 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.076113 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.076124 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.076141 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.076151 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.178389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.178434 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.178445 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.178461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.178475 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.281309 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.281341 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.281349 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.281361 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.281371 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.383947 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.383994 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.384005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.384022 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.384033 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.486550 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.486871 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.486880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.486897 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.486906 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.590315 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.590369 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.590380 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.590397 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.590411 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.692519 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.692598 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.692614 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.692635 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.692651 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.795785 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.795817 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.795826 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.795839 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.795848 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.899011 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.899059 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.899068 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.899082 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:08 crc kubenswrapper[5035]: I1002 09:29:08.899094 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:08Z","lastTransitionTime":"2025-10-02T09:29:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.001753 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.001790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.001799 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.001814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.001825 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.105146 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.105733 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.105930 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.106114 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.106293 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.162908 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.162944 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.162916 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.162908 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:09 crc kubenswrapper[5035]: E1002 09:29:09.163093 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:09 crc kubenswrapper[5035]: E1002 09:29:09.163385 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:09 crc kubenswrapper[5035]: E1002 09:29:09.163727 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:09 crc kubenswrapper[5035]: E1002 09:29:09.163796 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.208588 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.208639 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.208650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.208666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.208682 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.312357 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.312431 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.312447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.312473 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.312489 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.415041 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.415075 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.415083 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.415101 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.415118 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.517958 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.518018 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.518030 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.518049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.518064 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.620389 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.620435 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.620444 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.620461 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.620473 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.723425 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.723460 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.723469 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.723483 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.723491 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.826193 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.826228 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.826239 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.826254 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.826263 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.929802 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.929861 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.929872 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.929890 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:09 crc kubenswrapper[5035]: I1002 09:29:09.929904 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:09Z","lastTransitionTime":"2025-10-02T09:29:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.032105 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.032153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.032163 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.032179 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.032190 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.135065 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.135132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.135144 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.135162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.135174 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.163589 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:29:10 crc kubenswrapper[5035]: E1002 09:29:10.163756 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.237888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.237948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.237960 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.237975 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.237989 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.340797 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.340859 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.340871 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.340886 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.340993 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.443515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.443597 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.443615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.443637 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.443653 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.546511 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.546616 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.546643 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.546674 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.546696 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.649162 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.649273 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.649297 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.649326 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.649347 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.752062 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.752117 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.752132 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.752149 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.752163 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.854946 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.855004 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.855021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.855047 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.855063 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.957448 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.957506 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.957517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.957562 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:10 crc kubenswrapper[5035]: I1002 09:29:10.957574 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:10Z","lastTransitionTime":"2025-10-02T09:29:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.060615 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.060658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.060666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.060680 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.060689 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.161851 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:11 crc kubenswrapper[5035]: E1002 09:29:11.161947 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162102 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:11 crc kubenswrapper[5035]: E1002 09:29:11.162152 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162243 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:11 crc kubenswrapper[5035]: E1002 09:29:11.162279 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162311 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:11 crc kubenswrapper[5035]: E1002 09:29:11.162350 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162669 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162694 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162729 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162744 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.162755 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.265447 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.265571 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.265589 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.265610 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.265624 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.368383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.368436 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.368455 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.368479 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.368497 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.470987 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.471038 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.471050 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.471067 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.471079 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.573710 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.573798 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.573813 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.573829 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.573840 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.599555 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:11 crc kubenswrapper[5035]: E1002 09:29:11.599706 5035 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:29:11 crc kubenswrapper[5035]: E1002 09:29:11.599772 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs podName:452af00b-602d-43ab-a345-5453d6aebcf0 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:15.599754371 +0000 UTC m=+180.956098406 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs") pod "network-metrics-daemon-hzsjk" (UID: "452af00b-602d-43ab-a345-5453d6aebcf0") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.676498 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.676586 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.676650 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.676677 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.676717 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.778847 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.778891 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.778902 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.778919 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.778933 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.881805 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.881834 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.881843 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.881881 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.881889 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.984569 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.984607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.984617 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.984633 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:11 crc kubenswrapper[5035]: I1002 09:29:11.984645 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:11Z","lastTransitionTime":"2025-10-02T09:29:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.087727 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.087773 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.087785 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.087800 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.087814 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.190977 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.191021 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.191032 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.191048 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.191059 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.294841 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.294895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.294913 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.294934 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.294951 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.397474 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.397567 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.397585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.397608 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.397625 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.435888 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.435928 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.435937 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.435949 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.435959 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: E1002 09:29:12.452329 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.456658 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.456701 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.456716 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.456738 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.456754 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: E1002 09:29:12.468616 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.471880 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.471910 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.471925 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.471944 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.471957 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: E1002 09:29:12.494577 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.498787 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.498846 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.498868 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.498895 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.498916 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: E1002 09:29:12.518057 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.521693 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.521714 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.521722 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.521735 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.521744 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: E1002 09:29:12.545494 5035 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T09:29:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"44cc1aef-5d31-4451-bb6a-5114e2dcb900\\\",\\\"systemUUID\\\":\\\"a94415e2-ab3f-4649-975d-e580dfdd3e90\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T09:29:12Z is after 2025-08-24T17:21:41Z" Oct 02 09:29:12 crc kubenswrapper[5035]: E1002 09:29:12.545953 5035 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.556717 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.556771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.556789 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.556822 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.556839 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.659074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.659345 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.659422 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.659500 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.659607 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.762474 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.762523 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.762581 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.762604 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.762624 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.865061 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.865128 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.865140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.865157 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.865168 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.968140 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.968220 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.968235 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.968269 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:12 crc kubenswrapper[5035]: I1002 09:29:12.968288 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:12Z","lastTransitionTime":"2025-10-02T09:29:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.071268 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.071328 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.071343 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.071362 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.071374 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.162275 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:13 crc kubenswrapper[5035]: E1002 09:29:13.162648 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.162714 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.162717 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.162789 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:13 crc kubenswrapper[5035]: E1002 09:29:13.162829 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:13 crc kubenswrapper[5035]: E1002 09:29:13.162984 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:13 crc kubenswrapper[5035]: E1002 09:29:13.163045 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.174167 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.174222 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.174236 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.174256 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.174269 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.277517 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.277614 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.277636 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.277666 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.277694 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.380526 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.380607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.380620 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.380647 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.380678 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.483696 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.483776 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.483790 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.483808 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.483821 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.586669 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.586737 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.586754 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.586774 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.586788 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.689719 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.689774 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.689786 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.689806 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.689821 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.792697 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.792741 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.792753 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.792771 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.792784 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.896091 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.896153 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.896165 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.896182 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.896197 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.999419 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.999489 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.999513 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.999572 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:13 crc kubenswrapper[5035]: I1002 09:29:13.999600 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:13Z","lastTransitionTime":"2025-10-02T09:29:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.102889 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.102956 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.102977 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.103005 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.103029 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.205439 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.205521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.205592 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.205629 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.205653 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.308684 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.308733 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.308745 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.308765 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.308776 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.411475 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.411510 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.411518 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.411546 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.411555 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.514515 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.514599 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.514613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.514631 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.514644 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.617280 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.617344 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.617360 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.617382 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.617398 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.719764 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.719805 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.719814 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.719828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.719837 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.822503 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.822584 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.822596 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.822613 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.822622 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.926049 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.926095 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.926104 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.926118 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:14 crc kubenswrapper[5035]: I1002 09:29:14.926128 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:14Z","lastTransitionTime":"2025-10-02T09:29:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.029468 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.029521 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.029563 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.029585 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.029602 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.132137 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.132217 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.132238 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.132263 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.132282 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.162898 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.163020 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.163243 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.163244 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:15 crc kubenswrapper[5035]: E1002 09:29:15.163393 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:15 crc kubenswrapper[5035]: E1002 09:29:15.163526 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:15 crc kubenswrapper[5035]: E1002 09:29:15.163636 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:15 crc kubenswrapper[5035]: E1002 09:29:15.163754 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.235270 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.235302 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.235313 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.235329 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.235341 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.337758 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.337818 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.337828 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.337845 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.337855 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.440690 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.441227 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.441296 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.441388 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.441481 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.544327 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.544380 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.544394 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.544412 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.544424 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.646851 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.647155 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.647260 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.647383 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.647487 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.750023 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.750065 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.750074 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.750088 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.750098 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.852552 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.852593 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.852603 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.852618 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.852628 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.955427 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.955852 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.955948 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.956229 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:15 crc kubenswrapper[5035]: I1002 09:29:15.956289 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:15Z","lastTransitionTime":"2025-10-02T09:29:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.059003 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.059064 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.059081 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.059103 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.059118 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:16Z","lastTransitionTime":"2025-10-02T09:29:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:16 crc kubenswrapper[5035]: E1002 09:29:16.159731 5035 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.197559 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-5k5w7" podStartSLOduration=82.197540672 podStartE2EDuration="1m22.197540672s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.196038478 +0000 UTC m=+121.552382533" watchObservedRunningTime="2025-10-02 09:29:16.197540672 +0000 UTC m=+121.553884707" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.295506 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podStartSLOduration=83.295485034 podStartE2EDuration="1m23.295485034s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.262624454 +0000 UTC m=+121.618968499" watchObservedRunningTime="2025-10-02 09:29:16.295485034 +0000 UTC m=+121.651829079" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.306588 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-8xzpk" podStartSLOduration=82.306566891 podStartE2EDuration="1m22.306566891s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.306268953 +0000 UTC m=+121.662612988" watchObservedRunningTime="2025-10-02 09:29:16.306566891 +0000 UTC m=+121.662910926" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.323487 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=76.323470511 podStartE2EDuration="1m16.323470511s" podCreationTimestamp="2025-10-02 09:28:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.322748459 +0000 UTC m=+121.679092504" watchObservedRunningTime="2025-10-02 09:29:16.323470511 +0000 UTC m=+121.679814546" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.353110 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8rpfb" podStartSLOduration=83.353081875 podStartE2EDuration="1m23.353081875s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.352557079 +0000 UTC m=+121.708901114" watchObservedRunningTime="2025-10-02 09:29:16.353081875 +0000 UTC m=+121.709425920" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.364918 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-j6gm8" podStartSLOduration=83.364897064 podStartE2EDuration="1m23.364897064s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.364845042 +0000 UTC m=+121.721189097" watchObservedRunningTime="2025-10-02 09:29:16.364897064 +0000 UTC m=+121.721241089" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.380391 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=73.380373981 podStartE2EDuration="1m13.380373981s" podCreationTimestamp="2025-10-02 09:28:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.379389882 +0000 UTC m=+121.735733917" watchObservedRunningTime="2025-10-02 09:29:16.380373981 +0000 UTC m=+121.736718006" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.396450 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=55.396424125 podStartE2EDuration="55.396424125s" podCreationTimestamp="2025-10-02 09:28:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.395384104 +0000 UTC m=+121.751728139" watchObservedRunningTime="2025-10-02 09:29:16.396424125 +0000 UTC m=+121.752768150" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.470267 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=24.470238344 podStartE2EDuration="24.470238344s" podCreationTimestamp="2025-10-02 09:28:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.469392439 +0000 UTC m=+121.825736474" watchObservedRunningTime="2025-10-02 09:29:16.470238344 +0000 UTC m=+121.826582369" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.470493 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=31.470485201 podStartE2EDuration="31.470485201s" podCreationTimestamp="2025-10-02 09:28:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.440337171 +0000 UTC m=+121.796681186" watchObservedRunningTime="2025-10-02 09:29:16.470485201 +0000 UTC m=+121.826829246" Oct 02 09:29:16 crc kubenswrapper[5035]: I1002 09:29:16.486016 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-4hwst" podStartSLOduration=83.485996839 podStartE2EDuration="1m23.485996839s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:16.485760032 +0000 UTC m=+121.842104087" watchObservedRunningTime="2025-10-02 09:29:16.485996839 +0000 UTC m=+121.842340884" Oct 02 09:29:16 crc kubenswrapper[5035]: E1002 09:29:16.550796 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:17 crc kubenswrapper[5035]: I1002 09:29:17.162789 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:17 crc kubenswrapper[5035]: I1002 09:29:17.162850 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:17 crc kubenswrapper[5035]: E1002 09:29:17.163083 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:17 crc kubenswrapper[5035]: I1002 09:29:17.163109 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:17 crc kubenswrapper[5035]: I1002 09:29:17.163189 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:17 crc kubenswrapper[5035]: E1002 09:29:17.163289 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:17 crc kubenswrapper[5035]: E1002 09:29:17.163364 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:17 crc kubenswrapper[5035]: E1002 09:29:17.163450 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:19 crc kubenswrapper[5035]: I1002 09:29:19.162633 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:19 crc kubenswrapper[5035]: I1002 09:29:19.162715 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:19 crc kubenswrapper[5035]: I1002 09:29:19.162753 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:19 crc kubenswrapper[5035]: I1002 09:29:19.162768 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:19 crc kubenswrapper[5035]: E1002 09:29:19.164459 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:19 crc kubenswrapper[5035]: E1002 09:29:19.164638 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:19 crc kubenswrapper[5035]: E1002 09:29:19.164862 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:19 crc kubenswrapper[5035]: E1002 09:29:19.165058 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:21 crc kubenswrapper[5035]: I1002 09:29:21.162380 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:21 crc kubenswrapper[5035]: E1002 09:29:21.163576 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:21 crc kubenswrapper[5035]: I1002 09:29:21.162455 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:21 crc kubenswrapper[5035]: E1002 09:29:21.163966 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:21 crc kubenswrapper[5035]: I1002 09:29:21.162427 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:21 crc kubenswrapper[5035]: E1002 09:29:21.164201 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:21 crc kubenswrapper[5035]: I1002 09:29:21.162485 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:21 crc kubenswrapper[5035]: E1002 09:29:21.164450 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:21 crc kubenswrapper[5035]: E1002 09:29:21.552934 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.757412 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.757510 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.757559 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.757607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.757634 5035 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T09:29:22Z","lastTransitionTime":"2025-10-02T09:29:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.828275 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d"] Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.828986 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.831913 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.832276 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.832744 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.836021 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.924630 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d1463c5-de10-4e21-8580-febb257b5ce3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.925046 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0d1463c5-de10-4e21-8580-febb257b5ce3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.925239 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d1463c5-de10-4e21-8580-febb257b5ce3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.925406 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0d1463c5-de10-4e21-8580-febb257b5ce3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:22 crc kubenswrapper[5035]: I1002 09:29:22.925710 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0d1463c5-de10-4e21-8580-febb257b5ce3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.027355 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d1463c5-de10-4e21-8580-febb257b5ce3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.027447 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0d1463c5-de10-4e21-8580-febb257b5ce3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.027499 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d1463c5-de10-4e21-8580-febb257b5ce3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.027623 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0d1463c5-de10-4e21-8580-febb257b5ce3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.027703 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0d1463c5-de10-4e21-8580-febb257b5ce3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.028131 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0d1463c5-de10-4e21-8580-febb257b5ce3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.028171 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0d1463c5-de10-4e21-8580-febb257b5ce3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.029366 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0d1463c5-de10-4e21-8580-febb257b5ce3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.042736 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d1463c5-de10-4e21-8580-febb257b5ce3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.058251 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d1463c5-de10-4e21-8580-febb257b5ce3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-7922d\" (UID: \"0d1463c5-de10-4e21-8580-febb257b5ce3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.153337 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.162115 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:23 crc kubenswrapper[5035]: E1002 09:29:23.162396 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.162179 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.162249 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.162145 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:23 crc kubenswrapper[5035]: E1002 09:29:23.163158 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:23 crc kubenswrapper[5035]: E1002 09:29:23.163146 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:23 crc kubenswrapper[5035]: E1002 09:29:23.163004 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.163622 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:29:23 crc kubenswrapper[5035]: E1002 09:29:23.163888 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.596000 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" event={"ID":"0d1463c5-de10-4e21-8580-febb257b5ce3","Type":"ContainerStarted","Data":"d50c37364aaba6d6b85dd435812a6c55c1e8419f3f592863c4f5139d1864b34d"} Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.596328 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" event={"ID":"0d1463c5-de10-4e21-8580-febb257b5ce3","Type":"ContainerStarted","Data":"0d02cd755493f5724343adfa8968b92516a8e7c66bf5c4463637ac3f4ad46ffb"} Oct 02 09:29:23 crc kubenswrapper[5035]: I1002 09:29:23.616958 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-7922d" podStartSLOduration=90.616935979 podStartE2EDuration="1m30.616935979s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:23.61596282 +0000 UTC m=+128.972306875" watchObservedRunningTime="2025-10-02 09:29:23.616935979 +0000 UTC m=+128.973280004" Oct 02 09:29:25 crc kubenswrapper[5035]: I1002 09:29:25.162036 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:25 crc kubenswrapper[5035]: I1002 09:29:25.162135 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:25 crc kubenswrapper[5035]: I1002 09:29:25.162081 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:25 crc kubenswrapper[5035]: E1002 09:29:25.162205 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:25 crc kubenswrapper[5035]: E1002 09:29:25.162300 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:25 crc kubenswrapper[5035]: I1002 09:29:25.162317 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:25 crc kubenswrapper[5035]: E1002 09:29:25.162461 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:25 crc kubenswrapper[5035]: E1002 09:29:25.162637 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:26 crc kubenswrapper[5035]: E1002 09:29:26.553563 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:27 crc kubenswrapper[5035]: I1002 09:29:27.162665 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:27 crc kubenswrapper[5035]: I1002 09:29:27.162692 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:27 crc kubenswrapper[5035]: I1002 09:29:27.162787 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:27 crc kubenswrapper[5035]: E1002 09:29:27.162830 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:27 crc kubenswrapper[5035]: I1002 09:29:27.162686 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:27 crc kubenswrapper[5035]: E1002 09:29:27.163033 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:27 crc kubenswrapper[5035]: E1002 09:29:27.163168 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:27 crc kubenswrapper[5035]: E1002 09:29:27.163287 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:29 crc kubenswrapper[5035]: I1002 09:29:29.161948 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:29 crc kubenswrapper[5035]: I1002 09:29:29.161952 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:29 crc kubenswrapper[5035]: I1002 09:29:29.161983 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:29 crc kubenswrapper[5035]: E1002 09:29:29.162197 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:29 crc kubenswrapper[5035]: E1002 09:29:29.162273 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:29 crc kubenswrapper[5035]: E1002 09:29:29.162360 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:29 crc kubenswrapper[5035]: I1002 09:29:29.162827 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:29 crc kubenswrapper[5035]: E1002 09:29:29.163004 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:31 crc kubenswrapper[5035]: I1002 09:29:31.161978 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:31 crc kubenswrapper[5035]: I1002 09:29:31.162023 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:31 crc kubenswrapper[5035]: I1002 09:29:31.162064 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:31 crc kubenswrapper[5035]: E1002 09:29:31.162148 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:31 crc kubenswrapper[5035]: I1002 09:29:31.162181 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:31 crc kubenswrapper[5035]: E1002 09:29:31.162323 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:31 crc kubenswrapper[5035]: E1002 09:29:31.162510 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:31 crc kubenswrapper[5035]: E1002 09:29:31.162616 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:31 crc kubenswrapper[5035]: E1002 09:29:31.555369 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.162790 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.162968 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:33 crc kubenswrapper[5035]: E1002 09:29:33.163177 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.163308 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:33 crc kubenswrapper[5035]: E1002 09:29:33.163427 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:33 crc kubenswrapper[5035]: E1002 09:29:33.163485 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.163740 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:33 crc kubenswrapper[5035]: E1002 09:29:33.163914 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.635323 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/1.log" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.635902 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/0.log" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.635963 5035 generic.go:334] "Generic (PLEG): container finished" podID="5fafe5e6-83b2-447b-9379-b26e8071166b" containerID="bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b" exitCode=1 Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.636018 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerDied","Data":"bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b"} Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.636074 5035 scope.go:117] "RemoveContainer" containerID="d435a191f4dddd397a087963119a5829e452e8ea60642bc3dd21908b58e5a07e" Oct 02 09:29:33 crc kubenswrapper[5035]: I1002 09:29:33.637141 5035 scope.go:117] "RemoveContainer" containerID="bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b" Oct 02 09:29:33 crc kubenswrapper[5035]: E1002 09:29:33.637390 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8rpfb_openshift-multus(5fafe5e6-83b2-447b-9379-b26e8071166b)\"" pod="openshift-multus/multus-8rpfb" podUID="5fafe5e6-83b2-447b-9379-b26e8071166b" Oct 02 09:29:34 crc kubenswrapper[5035]: I1002 09:29:34.641792 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/1.log" Oct 02 09:29:35 crc kubenswrapper[5035]: I1002 09:29:35.162496 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:35 crc kubenswrapper[5035]: I1002 09:29:35.162496 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:35 crc kubenswrapper[5035]: I1002 09:29:35.162558 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:35 crc kubenswrapper[5035]: I1002 09:29:35.162590 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:35 crc kubenswrapper[5035]: E1002 09:29:35.163175 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:35 crc kubenswrapper[5035]: E1002 09:29:35.163411 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:35 crc kubenswrapper[5035]: E1002 09:29:35.163486 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:35 crc kubenswrapper[5035]: E1002 09:29:35.163625 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:35 crc kubenswrapper[5035]: I1002 09:29:35.163733 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:29:35 crc kubenswrapper[5035]: E1002 09:29:35.163999 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-cmx5n_openshift-ovn-kubernetes(ffeef997-af36-4315-8a12-46ecf69976e6)\"" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" Oct 02 09:29:36 crc kubenswrapper[5035]: E1002 09:29:36.555826 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:37 crc kubenswrapper[5035]: I1002 09:29:37.162099 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:37 crc kubenswrapper[5035]: I1002 09:29:37.162172 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:37 crc kubenswrapper[5035]: I1002 09:29:37.162244 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:37 crc kubenswrapper[5035]: E1002 09:29:37.162287 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:37 crc kubenswrapper[5035]: I1002 09:29:37.162103 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:37 crc kubenswrapper[5035]: E1002 09:29:37.162524 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:37 crc kubenswrapper[5035]: E1002 09:29:37.162702 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:37 crc kubenswrapper[5035]: E1002 09:29:37.162863 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:39 crc kubenswrapper[5035]: I1002 09:29:39.162454 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:39 crc kubenswrapper[5035]: I1002 09:29:39.162516 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:39 crc kubenswrapper[5035]: I1002 09:29:39.162585 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:39 crc kubenswrapper[5035]: E1002 09:29:39.162708 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:39 crc kubenswrapper[5035]: E1002 09:29:39.162823 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:39 crc kubenswrapper[5035]: E1002 09:29:39.162904 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:39 crc kubenswrapper[5035]: I1002 09:29:39.162489 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:39 crc kubenswrapper[5035]: E1002 09:29:39.163451 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:41 crc kubenswrapper[5035]: I1002 09:29:41.161983 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:41 crc kubenswrapper[5035]: I1002 09:29:41.162128 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:41 crc kubenswrapper[5035]: E1002 09:29:41.162160 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:41 crc kubenswrapper[5035]: I1002 09:29:41.162121 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:41 crc kubenswrapper[5035]: E1002 09:29:41.162300 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:41 crc kubenswrapper[5035]: I1002 09:29:41.162124 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:41 crc kubenswrapper[5035]: E1002 09:29:41.162455 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:41 crc kubenswrapper[5035]: E1002 09:29:41.162510 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:41 crc kubenswrapper[5035]: E1002 09:29:41.556960 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:43 crc kubenswrapper[5035]: I1002 09:29:43.162975 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:43 crc kubenswrapper[5035]: I1002 09:29:43.163014 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:43 crc kubenswrapper[5035]: I1002 09:29:43.163076 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:43 crc kubenswrapper[5035]: E1002 09:29:43.163129 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:43 crc kubenswrapper[5035]: I1002 09:29:43.163255 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:43 crc kubenswrapper[5035]: E1002 09:29:43.163458 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:43 crc kubenswrapper[5035]: E1002 09:29:43.163665 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:43 crc kubenswrapper[5035]: E1002 09:29:43.164122 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:45 crc kubenswrapper[5035]: I1002 09:29:45.162668 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:45 crc kubenswrapper[5035]: I1002 09:29:45.162749 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:45 crc kubenswrapper[5035]: I1002 09:29:45.162700 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:45 crc kubenswrapper[5035]: I1002 09:29:45.162892 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:45 crc kubenswrapper[5035]: E1002 09:29:45.162889 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:45 crc kubenswrapper[5035]: E1002 09:29:45.163075 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:45 crc kubenswrapper[5035]: E1002 09:29:45.163212 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:45 crc kubenswrapper[5035]: E1002 09:29:45.163509 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:46 crc kubenswrapper[5035]: E1002 09:29:46.557804 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:47 crc kubenswrapper[5035]: I1002 09:29:47.161976 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:47 crc kubenswrapper[5035]: I1002 09:29:47.162012 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:47 crc kubenswrapper[5035]: I1002 09:29:47.161992 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:47 crc kubenswrapper[5035]: I1002 09:29:47.162151 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:47 crc kubenswrapper[5035]: E1002 09:29:47.162130 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:47 crc kubenswrapper[5035]: E1002 09:29:47.162256 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:47 crc kubenswrapper[5035]: E1002 09:29:47.162450 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:47 crc kubenswrapper[5035]: E1002 09:29:47.162506 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:48 crc kubenswrapper[5035]: I1002 09:29:48.162743 5035 scope.go:117] "RemoveContainer" containerID="bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b" Oct 02 09:29:48 crc kubenswrapper[5035]: I1002 09:29:48.694826 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/1.log" Oct 02 09:29:48 crc kubenswrapper[5035]: I1002 09:29:48.694918 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerStarted","Data":"c84ee73fc0e719fd27b4324bcb86b267021b32fdb93655c846a1fd63b1c22189"} Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.162359 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.162399 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.162424 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:49 crc kubenswrapper[5035]: E1002 09:29:49.162682 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.162430 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:49 crc kubenswrapper[5035]: E1002 09:29:49.163000 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:49 crc kubenswrapper[5035]: E1002 09:29:49.163572 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:49 crc kubenswrapper[5035]: E1002 09:29:49.163619 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.164306 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.700800 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/3.log" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.704606 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerStarted","Data":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.705696 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.756626 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podStartSLOduration=115.75660839 podStartE2EDuration="1m55.75660839s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:29:49.75628493 +0000 UTC m=+155.112629025" watchObservedRunningTime="2025-10-02 09:29:49.75660839 +0000 UTC m=+155.112952425" Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.988412 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hzsjk"] Oct 02 09:29:49 crc kubenswrapper[5035]: I1002 09:29:49.988635 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:49 crc kubenswrapper[5035]: E1002 09:29:49.988819 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:51 crc kubenswrapper[5035]: I1002 09:29:51.162811 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:51 crc kubenswrapper[5035]: I1002 09:29:51.162880 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:51 crc kubenswrapper[5035]: E1002 09:29:51.162959 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:51 crc kubenswrapper[5035]: I1002 09:29:51.163012 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:51 crc kubenswrapper[5035]: E1002 09:29:51.163079 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:51 crc kubenswrapper[5035]: E1002 09:29:51.163330 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:51 crc kubenswrapper[5035]: E1002 09:29:51.559306 5035 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 09:29:52 crc kubenswrapper[5035]: I1002 09:29:52.162100 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:52 crc kubenswrapper[5035]: E1002 09:29:52.162332 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:53 crc kubenswrapper[5035]: I1002 09:29:53.162444 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:53 crc kubenswrapper[5035]: I1002 09:29:53.162514 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:53 crc kubenswrapper[5035]: I1002 09:29:53.162551 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:53 crc kubenswrapper[5035]: E1002 09:29:53.162690 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:53 crc kubenswrapper[5035]: E1002 09:29:53.162769 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:53 crc kubenswrapper[5035]: E1002 09:29:53.162823 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:54 crc kubenswrapper[5035]: I1002 09:29:54.162698 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:54 crc kubenswrapper[5035]: E1002 09:29:54.162937 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:55 crc kubenswrapper[5035]: I1002 09:29:55.163209 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:55 crc kubenswrapper[5035]: I1002 09:29:55.163299 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:55 crc kubenswrapper[5035]: I1002 09:29:55.163213 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:55 crc kubenswrapper[5035]: E1002 09:29:55.163428 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 09:29:55 crc kubenswrapper[5035]: E1002 09:29:55.163577 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 09:29:55 crc kubenswrapper[5035]: E1002 09:29:55.163666 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 09:29:55 crc kubenswrapper[5035]: I1002 09:29:55.578055 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:29:56 crc kubenswrapper[5035]: I1002 09:29:56.162744 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:56 crc kubenswrapper[5035]: E1002 09:29:56.163870 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hzsjk" podUID="452af00b-602d-43ab-a345-5453d6aebcf0" Oct 02 09:29:57 crc kubenswrapper[5035]: I1002 09:29:57.162736 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:29:57 crc kubenswrapper[5035]: I1002 09:29:57.163307 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:29:57 crc kubenswrapper[5035]: I1002 09:29:57.163634 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:29:57 crc kubenswrapper[5035]: I1002 09:29:57.165622 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 09:29:57 crc kubenswrapper[5035]: I1002 09:29:57.165821 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 09:29:57 crc kubenswrapper[5035]: I1002 09:29:57.166131 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 09:29:57 crc kubenswrapper[5035]: I1002 09:29:57.166810 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 09:29:58 crc kubenswrapper[5035]: I1002 09:29:58.162515 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:29:58 crc kubenswrapper[5035]: I1002 09:29:58.167730 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 09:29:58 crc kubenswrapper[5035]: I1002 09:29:58.167934 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.266282 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:03 crc kubenswrapper[5035]: E1002 09:30:03.266448 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:32:05.266429483 +0000 UTC m=+290.622773508 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.266519 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.275843 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.367381 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.367463 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.367520 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.368957 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.370452 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.371463 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.486247 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.502197 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.511663 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.751759 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2b6d0c95545d0b5137c7a61544a5a1b31f089ef0429c222d42bd579f2b20c470"} Oct 02 09:30:03 crc kubenswrapper[5035]: W1002 09:30:03.757389 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-b5cec6adaa9e670ac215e37420ff75b598e8f49777d4cd8cb99c711ad88f154a WatchSource:0}: Error finding container b5cec6adaa9e670ac215e37420ff75b598e8f49777d4cd8cb99c711ad88f154a: Status 404 returned error can't find the container with id b5cec6adaa9e670ac215e37420ff75b598e8f49777d4cd8cb99c711ad88f154a Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.898607 5035 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.950856 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-26669"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.951602 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qb4qh"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.951800 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.951932 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9qdrb"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.952438 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.953183 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.956932 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.957839 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.965459 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.965459 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.967216 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.967407 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.970057 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.982218 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.984998 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.985383 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.985710 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.985799 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.985880 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.986063 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.986180 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.986299 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.986428 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.986474 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.986564 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.986478 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.987145 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.987631 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.988095 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.988743 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.989157 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.989561 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.989904 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.990869 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.991068 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.991261 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.991431 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995492 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9975af5f-7bf6-41f9-919e-354f98ebc4d8-node-pullsecrets\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995563 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995600 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-config\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995623 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9975af5f-7bf6-41f9-919e-354f98ebc4d8-audit-dir\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995669 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-encryption-config\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995699 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-serving-cert\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995703 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995731 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-etcd-client\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995753 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-audit\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995773 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995794 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwpw8\" (UniqueName: \"kubernetes.io/projected/9975af5f-7bf6-41f9-919e-354f98ebc4d8-kube-api-access-wwpw8\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.995818 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-image-import-ca\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.997738 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-b7fjd"] Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.998214 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:03 crc kubenswrapper[5035]: I1002 09:30:03.999721 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.002440 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.002634 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.002782 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.003076 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.003129 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.003791 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.003972 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.004132 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.004378 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.006663 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.007098 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wgkqf"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.007358 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-96thd"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.007827 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.007842 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.008178 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: W1002 09:30:04.009707 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-274d2b9d6a551003e04da86ba7d1b697aa22b329585d71559183bc92babd588b WatchSource:0}: Error finding container 274d2b9d6a551003e04da86ba7d1b697aa22b329585d71559183bc92babd588b: Status 404 returned error can't find the container with id 274d2b9d6a551003e04da86ba7d1b697aa22b329585d71559183bc92babd588b Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.018119 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.018604 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.019231 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-m7jft"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.019735 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-m7jft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.020870 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.021022 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.021128 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.021250 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.021423 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.021563 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.031983 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fgjxt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.032913 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.033184 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.033406 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.033576 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.034011 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.034288 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.034368 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.034576 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.034810 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.034862 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.035039 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.035250 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pwdtp"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.035359 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.036134 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.036434 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7wsdb"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.036735 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.037468 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.051867 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.052079 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.052195 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.052309 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.052393 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.052470 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.052598 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.053021 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.053137 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.053252 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.053395 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.053704 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.053864 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.054040 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.054477 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.055064 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.055716 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.055814 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.056229 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.056439 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.056812 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.058269 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.058860 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.062031 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.062270 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.065238 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.084642 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.084785 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.084849 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.087054 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.087610 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.088097 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.088490 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.088729 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.088916 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.089083 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.089774 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.090336 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.092331 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2rf7w"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.092859 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096720 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-trusted-ca-bundle\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096762 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-audit-dir\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096792 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-encryption-config\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096841 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-serving-cert\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096869 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2284b\" (UniqueName: \"kubernetes.io/projected/ac14f12c-aed8-4802-80dd-18a7aee2f254-kube-api-access-2284b\") pod \"downloads-7954f5f757-m7jft\" (UID: \"ac14f12c-aed8-4802-80dd-18a7aee2f254\") " pod="openshift-console/downloads-7954f5f757-m7jft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096893 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qlkn\" (UniqueName: \"kubernetes.io/projected/44ea8dc6-acd0-4b46-b90e-78428e973475-kube-api-access-4qlkn\") pod \"cluster-samples-operator-665b6dd947-2nnft\" (UID: \"44ea8dc6-acd0-4b46-b90e-78428e973475\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096902 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.096914 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c3d581-243d-4d22-bf06-2771fec8192f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097099 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/368eaba6-c405-4ba6-afa9-13b32093851a-config\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097126 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/287782ba-f80b-4aae-a65a-9486f47a1e22-config\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097163 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097180 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwpw8\" (UniqueName: \"kubernetes.io/projected/9975af5f-7bf6-41f9-919e-354f98ebc4d8-kube-api-access-wwpw8\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097197 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk7qz\" (UniqueName: \"kubernetes.io/projected/96c3d581-243d-4d22-bf06-2771fec8192f-kube-api-access-kk7qz\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097216 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097236 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44ea8dc6-acd0-4b46-b90e-78428e973475-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2nnft\" (UID: \"44ea8dc6-acd0-4b46-b90e-78428e973475\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097251 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-service-ca-bundle\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097267 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-client-ca\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097291 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-image-import-ca\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097308 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f806291-b5e1-4b3c-b010-e0f00881a445-auth-proxy-config\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097322 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f806291-b5e1-4b3c-b010-e0f00881a445-config\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097339 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdth5\" (UniqueName: \"kubernetes.io/projected/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-kube-api-access-xdth5\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097358 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097374 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097388 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-config\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097403 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097418 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw6w9\" (UniqueName: \"kubernetes.io/projected/8abe751e-2b7d-4c51-ab84-904e88edbbbc-kube-api-access-kw6w9\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097432 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-oauth-config\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097448 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097466 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-config\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9975af5f-7bf6-41f9-919e-354f98ebc4d8-node-pullsecrets\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097488 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097501 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dec272a1-4b07-4d2d-99c5-776871d801ee-serving-cert\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097504 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097521 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097558 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097580 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0096104-afd1-4e4f-80f7-8e45211965f1-serving-cert\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097625 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-audit-policies\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097644 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2qvs\" (UniqueName: \"kubernetes.io/projected/368eaba6-c405-4ba6-afa9-13b32093851a-kube-api-access-w2qvs\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097664 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097685 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-config\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097702 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc9lq\" (UniqueName: \"kubernetes.io/projected/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-kube-api-access-rc9lq\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097715 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098044 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098160 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098237 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098357 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098410 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-etcd-serving-ca\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098449 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098550 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098559 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098677 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098726 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097720 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/368eaba6-c405-4ba6-afa9-13b32093851a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098834 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/287782ba-f80b-4aae-a65a-9486f47a1e22-serving-cert\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098863 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-encryption-config\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098892 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvlhw\" (UniqueName: \"kubernetes.io/projected/f0096104-afd1-4e4f-80f7-8e45211965f1-kube-api-access-mvlhw\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098921 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-config\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098942 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9975af5f-7bf6-41f9-919e-354f98ebc4d8-audit-dir\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098965 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-client-ca\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098987 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-service-ca\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099009 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.097500 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099037 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-serving-cert\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099061 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jqwf\" (UniqueName: \"kubernetes.io/projected/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-kube-api-access-6jqwf\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099085 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099108 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-policies\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099133 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f806291-b5e1-4b3c-b010-e0f00881a445-machine-approver-tls\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099156 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-etcd-client\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099179 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-oauth-serving-cert\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099205 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxgbv\" (UniqueName: \"kubernetes.io/projected/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-kube-api-access-pxgbv\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099229 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-serving-cert\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099250 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099269 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099293 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22z4q\" (UniqueName: \"kubernetes.io/projected/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-kube-api-access-22z4q\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099317 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-audit\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099341 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-etcd-client\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099375 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100076 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-config\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100106 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-dir\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100130 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p48dj\" (UniqueName: \"kubernetes.io/projected/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-kube-api-access-p48dj\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100155 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100180 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100200 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100221 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f0096104-afd1-4e4f-80f7-8e45211965f1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100242 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-config\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100265 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100344 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9975af5f-7bf6-41f9-919e-354f98ebc4d8-node-pullsecrets\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100376 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzppv\" (UniqueName: \"kubernetes.io/projected/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-kube-api-access-dzppv\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100404 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-client\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100432 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100450 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-ca\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100464 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100688 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100782 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100908 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101050 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101718 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.100471 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdw8j\" (UniqueName: \"kubernetes.io/projected/dec272a1-4b07-4d2d-99c5-776871d801ee-kube-api-access-cdw8j\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101792 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-service-ca\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101811 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctrzx\" (UniqueName: \"kubernetes.io/projected/5943cfb7-f43b-4a5f-8ac2-b39e8e898c56-kube-api-access-ctrzx\") pod \"dns-operator-744455d44c-96thd\" (UID: \"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56\") " pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101829 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-metrics-tls\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101853 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-trusted-ca\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101870 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/287782ba-f80b-4aae-a65a-9486f47a1e22-trusted-ca\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101888 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-serving-cert\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101935 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-serving-cert\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101952 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tsq7\" (UniqueName: \"kubernetes.io/projected/8f806291-b5e1-4b3c-b010-e0f00881a445-kube-api-access-2tsq7\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101969 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5943cfb7-f43b-4a5f-8ac2-b39e8e898c56-metrics-tls\") pod \"dns-operator-744455d44c-96thd\" (UID: \"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56\") " pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.101993 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-serving-cert\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102023 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/368eaba6-c405-4ba6-afa9-13b32093851a-images\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102041 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96c3d581-243d-4d22-bf06-2771fec8192f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102058 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6545\" (UniqueName: \"kubernetes.io/projected/287782ba-f80b-4aae-a65a-9486f47a1e22-kube-api-access-w6545\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102082 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102192 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102218 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-encryption-config\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102255 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102323 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102399 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.098200 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.102891 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.099316 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-image-import-ca\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.103059 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.103074 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-rnzld"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.103117 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.103183 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-config\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.103278 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9975af5f-7bf6-41f9-919e-354f98ebc4d8-audit-dir\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.103692 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.104827 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.105245 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.105973 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9975af5f-7bf6-41f9-919e-354f98ebc4d8-audit\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.107925 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-serving-cert\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.107944 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9975af5f-7bf6-41f9-919e-354f98ebc4d8-etcd-client\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.107979 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.108636 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.112564 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qb4qh"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.113313 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.116984 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.117973 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.118656 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.121350 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.122748 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9qdrb"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.123649 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-b7fjd"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.125141 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.126136 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-26669"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.127775 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.130037 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.130251 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.131798 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.132691 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.134006 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.134518 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-dcmtw"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.136488 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.136593 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.142117 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jbph8"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.142306 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.143043 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4ggtp"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.144143 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.145499 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.149199 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.152083 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.153378 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.154585 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.160658 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.161011 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.162731 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.162768 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xshqj"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.163082 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.164510 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.171351 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-44nt7"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.171847 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wgkqf"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.171921 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.172497 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.173926 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.174655 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.176508 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.178977 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-m7jft"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.181689 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.182688 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7wsdb"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.183732 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-96thd"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.184834 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.186074 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dcmtw"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.187344 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fgjxt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.188426 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.189385 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.190487 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.191590 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.192876 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.193828 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.194688 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.196922 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2rf7w"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.198094 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.199264 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.200362 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jbph8"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.201423 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4ggtp"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202630 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-stats-auth\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202695 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdw8j\" (UniqueName: \"kubernetes.io/projected/dec272a1-4b07-4d2d-99c5-776871d801ee-kube-api-access-cdw8j\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202719 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/287782ba-f80b-4aae-a65a-9486f47a1e22-trusted-ca\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202734 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-trusted-ca\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202751 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-serving-cert\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202769 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202785 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/42daaa36-95f5-4fd6-bcbd-842b290ab83a-proxy-tls\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202790 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202799 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-serving-cert\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202874 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8336420c-cced-4dc7-bd94-e95ae46cdbc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hwsxs\" (UID: \"8336420c-cced-4dc7-bd94-e95ae46cdbc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202909 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96c3d581-243d-4d22-bf06-2771fec8192f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202933 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-trusted-ca-bundle\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202952 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-config\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202970 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2284b\" (UniqueName: \"kubernetes.io/projected/ac14f12c-aed8-4802-80dd-18a7aee2f254-kube-api-access-2284b\") pod \"downloads-7954f5f757-m7jft\" (UID: \"ac14f12c-aed8-4802-80dd-18a7aee2f254\") " pod="openshift-console/downloads-7954f5f757-m7jft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.202987 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qlkn\" (UniqueName: \"kubernetes.io/projected/44ea8dc6-acd0-4b46-b90e-78428e973475-kube-api-access-4qlkn\") pod \"cluster-samples-operator-665b6dd947-2nnft\" (UID: \"44ea8dc6-acd0-4b46-b90e-78428e973475\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203003 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c3d581-243d-4d22-bf06-2771fec8192f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203030 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203059 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk7qz\" (UniqueName: \"kubernetes.io/projected/96c3d581-243d-4d22-bf06-2771fec8192f-kube-api-access-kk7qz\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203078 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203095 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-client-ca\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203115 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-config\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203133 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203149 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw6w9\" (UniqueName: \"kubernetes.io/projected/8abe751e-2b7d-4c51-ab84-904e88edbbbc-kube-api-access-kw6w9\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203165 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203184 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-config\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203201 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf75h\" (UniqueName: \"kubernetes.io/projected/4fb71e4c-4c84-482f-819b-c32ba30776ec-kube-api-access-lf75h\") pod \"migrator-59844c95c7-dq99k\" (UID: \"4fb71e4c-4c84-482f-819b-c32ba30776ec\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203219 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd6wk\" (UniqueName: \"kubernetes.io/projected/e9af7278-65f4-471e-8b74-4305e9174038-kube-api-access-pd6wk\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203239 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1b15a92-86a9-4ab9-89d6-5de58811c55c-proxy-tls\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203263 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203290 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203310 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42daaa36-95f5-4fd6-bcbd-842b290ab83a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203331 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d4px\" (UniqueName: \"kubernetes.io/projected/3e222bb5-92b3-4a4d-833a-d383bdcc4515-kube-api-access-8d4px\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203357 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203379 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc9lq\" (UniqueName: \"kubernetes.io/projected/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-kube-api-access-rc9lq\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203402 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/368eaba6-c405-4ba6-afa9-13b32093851a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203427 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-service-ca\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203452 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-srv-cert\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203476 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4ggtp\" (UID: \"86e0567f-4ddd-4034-8c0f-a9e7c009de8c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203498 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-etcd-client\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203521 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-serving-cert\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203565 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203586 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-policies\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203608 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f806291-b5e1-4b3c-b010-e0f00881a445-machine-approver-tls\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203632 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1717ec80-35f4-4622-8b21-abdf91627e98-cert\") pod \"ingress-canary-dcmtw\" (UID: \"1717ec80-35f4-4622-8b21-abdf91627e98\") " pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203654 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxgbv\" (UniqueName: \"kubernetes.io/projected/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-kube-api-access-pxgbv\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203676 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-serving-cert\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203698 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203721 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22z4q\" (UniqueName: \"kubernetes.io/projected/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-kube-api-access-22z4q\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203745 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25ng6\" (UniqueName: \"kubernetes.io/projected/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-kube-api-access-25ng6\") pod \"multus-admission-controller-857f4d67dd-4ggtp\" (UID: \"86e0567f-4ddd-4034-8c0f-a9e7c009de8c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203770 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndr5t\" (UniqueName: \"kubernetes.io/projected/977e3b39-af15-4940-ba0d-1e22466018c8-kube-api-access-ndr5t\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203793 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-dir\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203813 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p48dj\" (UniqueName: \"kubernetes.io/projected/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-kube-api-access-p48dj\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203832 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1b15a92-86a9-4ab9-89d6-5de58811c55c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203842 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pwdtp"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203851 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/287782ba-f80b-4aae-a65a-9486f47a1e22-trusted-ca\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203909 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-default-certificate\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.203940 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzt7g\" (UniqueName: \"kubernetes.io/projected/f1b15a92-86a9-4ab9-89d6-5de58811c55c-kube-api-access-rzt7g\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204019 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-config\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204046 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204070 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzppv\" (UniqueName: \"kubernetes.io/projected/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-kube-api-access-dzppv\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204097 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-client\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204121 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-profile-collector-cert\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204145 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-ca\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204166 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-service-ca\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204183 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-trusted-ca-bundle\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204188 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctrzx\" (UniqueName: \"kubernetes.io/projected/5943cfb7-f43b-4a5f-8ac2-b39e8e898c56-kube-api-access-ctrzx\") pod \"dns-operator-744455d44c-96thd\" (UID: \"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56\") " pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204240 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-metrics-tls\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204265 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204291 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204324 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-config\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204584 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.204947 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.205615 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-service-ca\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.205998 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-config\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.206098 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-dir\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.206525 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96c3d581-243d-4d22-bf06-2771fec8192f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.206921 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-serving-cert\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.206955 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/96c3d581-243d-4d22-bf06-2771fec8192f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.207060 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.207093 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-serving-cert\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.207585 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-serving-cert\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.207643 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.207687 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-client-ca\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.207803 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.208102 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tsq7\" (UniqueName: \"kubernetes.io/projected/8f806291-b5e1-4b3c-b010-e0f00881a445-kube-api-access-2tsq7\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.208161 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5943cfb7-f43b-4a5f-8ac2-b39e8e898c56-metrics-tls\") pod \"dns-operator-744455d44c-96thd\" (UID: \"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56\") " pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.208190 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/368eaba6-c405-4ba6-afa9-13b32093851a-images\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.208335 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6545\" (UniqueName: \"kubernetes.io/projected/287782ba-f80b-4aae-a65a-9486f47a1e22-kube-api-access-w6545\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.208359 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.208617 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-config\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209062 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-serving-cert\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209116 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-config\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209128 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/368eaba6-c405-4ba6-afa9-13b32093851a-images\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209168 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-audit-dir\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209203 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-audit-dir\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209244 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc4gd\" (UniqueName: \"kubernetes.io/projected/1717ec80-35f4-4622-8b21-abdf91627e98-kube-api-access-zc4gd\") pod \"ingress-canary-dcmtw\" (UID: \"1717ec80-35f4-4622-8b21-abdf91627e98\") " pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209303 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209329 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-service-ca\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209337 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209331 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4r6p\" (UniqueName: \"kubernetes.io/projected/8336420c-cced-4dc7-bd94-e95ae46cdbc4-kube-api-access-m4r6p\") pod \"package-server-manager-789f6589d5-hwsxs\" (UID: \"8336420c-cced-4dc7-bd94-e95ae46cdbc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209383 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/368eaba6-c405-4ba6-afa9-13b32093851a-config\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209451 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/287782ba-f80b-4aae-a65a-9486f47a1e22-config\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209492 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209609 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hphx\" (UniqueName: \"kubernetes.io/projected/10d07013-d746-4fc0-8485-64d95703dd12-kube-api-access-8hphx\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209636 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqjj5\" (UniqueName: \"kubernetes.io/projected/42daaa36-95f5-4fd6-bcbd-842b290ab83a-kube-api-access-jqjj5\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209653 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-metrics-certs\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209681 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44ea8dc6-acd0-4b46-b90e-78428e973475-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2nnft\" (UID: \"44ea8dc6-acd0-4b46-b90e-78428e973475\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209710 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-service-ca-bundle\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209723 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-ca\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.209968 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210039 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdth5\" (UniqueName: \"kubernetes.io/projected/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-kube-api-access-xdth5\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210065 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/368eaba6-c405-4ba6-afa9-13b32093851a-config\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210076 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210084 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f806291-b5e1-4b3c-b010-e0f00881a445-auth-proxy-config\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210170 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f806291-b5e1-4b3c-b010-e0f00881a445-config\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210319 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210343 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/287782ba-f80b-4aae-a65a-9486f47a1e22-config\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210379 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-service-ca-bundle\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210394 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-oauth-config\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210458 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dec272a1-4b07-4d2d-99c5-776871d801ee-serving-cert\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210488 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-audit-policies\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210558 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0096104-afd1-4e4f-80f7-8e45211965f1-serving-cert\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210578 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f806291-b5e1-4b3c-b010-e0f00881a445-config\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210591 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9af7278-65f4-471e-8b74-4305e9174038-config-volume\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210622 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2qvs\" (UniqueName: \"kubernetes.io/projected/368eaba6-c405-4ba6-afa9-13b32093851a-kube-api-access-w2qvs\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210651 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-config\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210678 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e222bb5-92b3-4a4d-833a-d383bdcc4515-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210839 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8f806291-b5e1-4b3c-b010-e0f00881a445-auth-proxy-config\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210891 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/287782ba-f80b-4aae-a65a-9486f47a1e22-serving-cert\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210917 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-encryption-config\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210944 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvlhw\" (UniqueName: \"kubernetes.io/projected/f0096104-afd1-4e4f-80f7-8e45211965f1-kube-api-access-mvlhw\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.210969 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9af7278-65f4-471e-8b74-4305e9174038-secret-volume\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211044 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e222bb5-92b3-4a4d-833a-d383bdcc4515-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211076 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-client-ca\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211102 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211143 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jqwf\" (UniqueName: \"kubernetes.io/projected/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-kube-api-access-6jqwf\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211214 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-oauth-serving-cert\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211246 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e222bb5-92b3-4a4d-833a-d383bdcc4515-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211272 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211294 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/977e3b39-af15-4940-ba0d-1e22466018c8-service-ca-bundle\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211323 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211346 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42daaa36-95f5-4fd6-bcbd-842b290ab83a-images\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211371 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-config\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211389 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-config\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.211787 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-etcd-client\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.212217 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-etcd-client\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.212647 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-serving-cert\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.212647 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-audit-policies\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.212915 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.212959 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xshqj"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.212987 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-client-ca\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.213017 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44ea8dc6-acd0-4b46-b90e-78428e973475-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2nnft\" (UID: \"44ea8dc6-acd0-4b46-b90e-78428e973475\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.213098 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.213177 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-oauth-config\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.213356 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.213454 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.213502 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f0096104-afd1-4e4f-80f7-8e45211965f1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.213937 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f0096104-afd1-4e4f-80f7-8e45211965f1-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.214296 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-oauth-serving-cert\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.215115 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-encryption-config\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.215177 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.215308 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0096104-afd1-4e4f-80f7-8e45211965f1-serving-cert\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.215365 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.215612 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dec272a1-4b07-4d2d-99c5-776871d801ee-config\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.216041 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/287782ba-f80b-4aae-a65a-9486f47a1e22-serving-cert\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.216265 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.216450 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-metrics-tls\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.216522 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5943cfb7-f43b-4a5f-8ac2-b39e8e898c56-metrics-tls\") pod \"dns-operator-744455d44c-96thd\" (UID: \"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56\") " pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.216981 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8f806291-b5e1-4b3c-b010-e0f00881a445-machine-approver-tls\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.217212 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-serving-cert\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.217396 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.217396 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/368eaba6-c405-4ba6-afa9-13b32093851a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.217685 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gpk2c"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.218843 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.218919 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.219221 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.219258 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dec272a1-4b07-4d2d-99c5-776871d801ee-serving-cert\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.220148 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6cktt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.220842 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.221329 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.221588 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.228289 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gpk2c"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.231502 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6cktt"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.233101 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-44nt7"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.234079 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-6kjww"] Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.234833 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.245873 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.250372 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.258467 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.269305 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.279361 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.284571 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.294640 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.314913 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.315622 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.315923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.316166 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1b15a92-86a9-4ab9-89d6-5de58811c55c-proxy-tls\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.316314 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf75h\" (UniqueName: \"kubernetes.io/projected/4fb71e4c-4c84-482f-819b-c32ba30776ec-kube-api-access-lf75h\") pod \"migrator-59844c95c7-dq99k\" (UID: \"4fb71e4c-4c84-482f-819b-c32ba30776ec\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.316352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd6wk\" (UniqueName: \"kubernetes.io/projected/e9af7278-65f4-471e-8b74-4305e9174038-kube-api-access-pd6wk\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.316504 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42daaa36-95f5-4fd6-bcbd-842b290ab83a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.316580 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d4px\" (UniqueName: \"kubernetes.io/projected/3e222bb5-92b3-4a4d-833a-d383bdcc4515-kube-api-access-8d4px\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.316642 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-srv-cert\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.316966 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4ggtp\" (UID: \"86e0567f-4ddd-4034-8c0f-a9e7c009de8c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317063 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1717ec80-35f4-4622-8b21-abdf91627e98-cert\") pod \"ingress-canary-dcmtw\" (UID: \"1717ec80-35f4-4622-8b21-abdf91627e98\") " pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317154 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25ng6\" (UniqueName: \"kubernetes.io/projected/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-kube-api-access-25ng6\") pod \"multus-admission-controller-857f4d67dd-4ggtp\" (UID: \"86e0567f-4ddd-4034-8c0f-a9e7c009de8c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317248 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndr5t\" (UniqueName: \"kubernetes.io/projected/977e3b39-af15-4940-ba0d-1e22466018c8-kube-api-access-ndr5t\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317320 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/42daaa36-95f5-4fd6-bcbd-842b290ab83a-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317337 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1b15a92-86a9-4ab9-89d6-5de58811c55c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317482 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-default-certificate\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317512 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzt7g\" (UniqueName: \"kubernetes.io/projected/f1b15a92-86a9-4ab9-89d6-5de58811c55c-kube-api-access-rzt7g\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317735 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-profile-collector-cert\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317764 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317910 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-config\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317934 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.317975 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318088 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc4gd\" (UniqueName: \"kubernetes.io/projected/1717ec80-35f4-4622-8b21-abdf91627e98-kube-api-access-zc4gd\") pod \"ingress-canary-dcmtw\" (UID: \"1717ec80-35f4-4622-8b21-abdf91627e98\") " pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318110 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318165 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4r6p\" (UniqueName: \"kubernetes.io/projected/8336420c-cced-4dc7-bd94-e95ae46cdbc4-kube-api-access-m4r6p\") pod \"package-server-manager-789f6589d5-hwsxs\" (UID: \"8336420c-cced-4dc7-bd94-e95ae46cdbc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318217 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hphx\" (UniqueName: \"kubernetes.io/projected/10d07013-d746-4fc0-8485-64d95703dd12-kube-api-access-8hphx\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqjj5\" (UniqueName: \"kubernetes.io/projected/42daaa36-95f5-4fd6-bcbd-842b290ab83a-kube-api-access-jqjj5\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318287 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-metrics-certs\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318389 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9af7278-65f4-471e-8b74-4305e9174038-config-volume\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318459 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9af7278-65f4-471e-8b74-4305e9174038-secret-volume\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318668 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e222bb5-92b3-4a4d-833a-d383bdcc4515-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318711 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e222bb5-92b3-4a4d-833a-d383bdcc4515-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318757 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e222bb5-92b3-4a4d-833a-d383bdcc4515-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318775 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/977e3b39-af15-4940-ba0d-1e22466018c8-service-ca-bundle\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318814 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42daaa36-95f5-4fd6-bcbd-842b290ab83a-images\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318838 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-stats-auth\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318889 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318910 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/42daaa36-95f5-4fd6-bcbd-842b290ab83a-proxy-tls\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318948 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8336420c-cced-4dc7-bd94-e95ae46cdbc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hwsxs\" (UID: \"8336420c-cced-4dc7-bd94-e95ae46cdbc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.318987 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-config\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.320169 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1b15a92-86a9-4ab9-89d6-5de58811c55c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.334047 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.338809 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-policies\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.354598 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.364451 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.381008 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.384781 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-trusted-ca\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.393880 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.414311 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.434082 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.441160 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1b15a92-86a9-4ab9-89d6-5de58811c55c-proxy-tls\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.453860 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.461021 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42daaa36-95f5-4fd6-bcbd-842b290ab83a-images\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.473787 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.495493 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.504228 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.514836 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.520044 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-config\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.534282 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.555164 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.560947 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.574518 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.580164 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.594664 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.613846 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.635630 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.644278 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/42daaa36-95f5-4fd6-bcbd-842b290ab83a-proxy-tls\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.655511 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.663821 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.675114 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.694210 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.700627 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-config\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.714104 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.743648 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.752045 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e222bb5-92b3-4a4d-833a-d383bdcc4515-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.754300 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.757777 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0e80d56ae926ba4921be4ba3baeef75c0ff6af34eeea9cc4f1e01f82bef66147"} Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.759119 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2a10ab2f86d385681b309277f67f3bec87b586b9a7126b4a247cf1512aa87997"} Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.759168 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"274d2b9d6a551003e04da86ba7d1b697aa22b329585d71559183bc92babd588b"} Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.759379 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.760586 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"dc316a92dfc38acd347224dcb0e2fc5d0283d2aa286e29eea151f13a4bc604d7"} Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.760644 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"b5cec6adaa9e670ac215e37420ff75b598e8f49777d4cd8cb99c711ad88f154a"} Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.774181 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.794855 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.828140 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwpw8\" (UniqueName: \"kubernetes.io/projected/9975af5f-7bf6-41f9-919e-354f98ebc4d8-kube-api-access-wwpw8\") pod \"apiserver-76f77b778f-26669\" (UID: \"9975af5f-7bf6-41f9-919e-354f98ebc4d8\") " pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.854680 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.863433 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e222bb5-92b3-4a4d-833a-d383bdcc4515-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.873961 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.894242 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.903449 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.914312 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.924726 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-metrics-certs\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.935650 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.959913 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.967321 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-default-certificate\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.974876 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.984938 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/977e3b39-af15-4940-ba0d-1e22466018c8-stats-auth\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:04 crc kubenswrapper[5035]: I1002 09:30:04.999517 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.011085 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/977e3b39-af15-4940-ba0d-1e22466018c8-service-ca-bundle\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.016728 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.035764 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.053913 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.074014 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.094884 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.115101 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.133545 5035 request.go:700] Waited for 1.003237838s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dkube-storage-version-migrator-operator-dockercfg-2bh8d&limit=500&resourceVersion=0 Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.135473 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.155232 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.160085 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-26669"] Oct 02 09:30:05 crc kubenswrapper[5035]: W1002 09:30:05.165093 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9975af5f_7bf6_41f9_919e_354f98ebc4d8.slice/crio-ab08221f1d00de89ca711f51713b795079f1db41f847c3fd6a03d1716e250fd5 WatchSource:0}: Error finding container ab08221f1d00de89ca711f51713b795079f1db41f847c3fd6a03d1716e250fd5: Status 404 returned error can't find the container with id ab08221f1d00de89ca711f51713b795079f1db41f847c3fd6a03d1716e250fd5 Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.174192 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.195027 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.215093 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.234483 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.255314 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.277039 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.283000 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-profile-collector-cert\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.283440 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9af7278-65f4-471e-8b74-4305e9174038-secret-volume\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.296235 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.299843 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9af7278-65f4-471e-8b74-4305e9174038-config-volume\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.314823 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.317272 5035 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.317389 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-srv-cert podName:10d07013-d746-4fc0-8485-64d95703dd12 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:05.817351545 +0000 UTC m=+171.173695630 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-srv-cert") pod "catalog-operator-68c6474976-ddmmt" (UID: "10d07013-d746-4fc0-8485-64d95703dd12") : failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.318721 5035 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.318796 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-webhook-certs podName:86e0567f-4ddd-4034-8c0f-a9e7c009de8c nodeName:}" failed. No retries permitted until 2025-10-02 09:30:05.818775487 +0000 UTC m=+171.175119552 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-webhook-certs") pod "multus-admission-controller-857f4d67dd-4ggtp" (UID: "86e0567f-4ddd-4034-8c0f-a9e7c009de8c") : failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.318842 5035 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.318904 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1717ec80-35f4-4622-8b21-abdf91627e98-cert podName:1717ec80-35f4-4622-8b21-abdf91627e98 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:05.81888484 +0000 UTC m=+171.175228905 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1717ec80-35f4-4622-8b21-abdf91627e98-cert") pod "ingress-canary-dcmtw" (UID: "1717ec80-35f4-4622-8b21-abdf91627e98") : failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.320628 5035 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: E1002 09:30:05.321356 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8336420c-cced-4dc7-bd94-e95ae46cdbc4-package-server-manager-serving-cert podName:8336420c-cced-4dc7-bd94-e95ae46cdbc4 nodeName:}" failed. No retries permitted until 2025-10-02 09:30:05.821336591 +0000 UTC m=+171.177680616 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8336420c-cced-4dc7-bd94-e95ae46cdbc4-package-server-manager-serving-cert") pod "package-server-manager-789f6589d5-hwsxs" (UID: "8336420c-cced-4dc7-bd94-e95ae46cdbc4") : failed to sync secret cache: timed out waiting for the condition Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.334010 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.354100 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.374623 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.394197 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.414361 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.437004 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.454710 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.477072 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.494183 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.513247 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.542024 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.553919 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.575396 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.594503 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.613987 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.634895 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.655361 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.674915 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.694087 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.715359 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.733886 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.765277 5035 generic.go:334] "Generic (PLEG): container finished" podID="9975af5f-7bf6-41f9-919e-354f98ebc4d8" containerID="3fde717e423571911fbd542dff0b13e2fa35fafbcdd925a441de1d62e9da41d2" exitCode=0 Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.765362 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-26669" event={"ID":"9975af5f-7bf6-41f9-919e-354f98ebc4d8","Type":"ContainerDied","Data":"3fde717e423571911fbd542dff0b13e2fa35fafbcdd925a441de1d62e9da41d2"} Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.765424 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-26669" event={"ID":"9975af5f-7bf6-41f9-919e-354f98ebc4d8","Type":"ContainerStarted","Data":"ab08221f1d00de89ca711f51713b795079f1db41f847c3fd6a03d1716e250fd5"} Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.774968 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.794009 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.814851 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.835598 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.839333 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8336420c-cced-4dc7-bd94-e95ae46cdbc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hwsxs\" (UID: \"8336420c-cced-4dc7-bd94-e95ae46cdbc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.839518 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-srv-cert\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.839622 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4ggtp\" (UID: \"86e0567f-4ddd-4034-8c0f-a9e7c009de8c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.839674 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1717ec80-35f4-4622-8b21-abdf91627e98-cert\") pod \"ingress-canary-dcmtw\" (UID: \"1717ec80-35f4-4622-8b21-abdf91627e98\") " pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.846861 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4ggtp\" (UID: \"86e0567f-4ddd-4034-8c0f-a9e7c009de8c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.848133 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1717ec80-35f4-4622-8b21-abdf91627e98-cert\") pod \"ingress-canary-dcmtw\" (UID: \"1717ec80-35f4-4622-8b21-abdf91627e98\") " pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.852163 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8336420c-cced-4dc7-bd94-e95ae46cdbc4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hwsxs\" (UID: \"8336420c-cced-4dc7-bd94-e95ae46cdbc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.852896 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/10d07013-d746-4fc0-8485-64d95703dd12-srv-cert\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.853976 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.894443 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdw8j\" (UniqueName: \"kubernetes.io/projected/dec272a1-4b07-4d2d-99c5-776871d801ee-kube-api-access-cdw8j\") pod \"authentication-operator-69f744f599-pwdtp\" (UID: \"dec272a1-4b07-4d2d-99c5-776871d801ee\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.912294 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qlkn\" (UniqueName: \"kubernetes.io/projected/44ea8dc6-acd0-4b46-b90e-78428e973475-kube-api-access-4qlkn\") pod \"cluster-samples-operator-665b6dd947-2nnft\" (UID: \"44ea8dc6-acd0-4b46-b90e-78428e973475\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.921326 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.931708 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2284b\" (UniqueName: \"kubernetes.io/projected/ac14f12c-aed8-4802-80dd-18a7aee2f254-kube-api-access-2284b\") pod \"downloads-7954f5f757-m7jft\" (UID: \"ac14f12c-aed8-4802-80dd-18a7aee2f254\") " pod="openshift-console/downloads-7954f5f757-m7jft" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.959329 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctrzx\" (UniqueName: \"kubernetes.io/projected/5943cfb7-f43b-4a5f-8ac2-b39e8e898c56-kube-api-access-ctrzx\") pod \"dns-operator-744455d44c-96thd\" (UID: \"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56\") " pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.977323 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22z4q\" (UniqueName: \"kubernetes.io/projected/5ff717d1-2a4c-4f7f-a80b-c043a016d8ca-kube-api-access-22z4q\") pod \"openshift-apiserver-operator-796bbdcf4f-wxrhf\" (UID: \"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.984360 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-m7jft" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.997270 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-bound-sa-token\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:05 crc kubenswrapper[5035]: I1002 09:30:05.997573 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.011827 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxgbv\" (UniqueName: \"kubernetes.io/projected/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-kube-api-access-pxgbv\") pod \"route-controller-manager-6576b87f9c-ldpj6\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.030919 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw6w9\" (UniqueName: \"kubernetes.io/projected/8abe751e-2b7d-4c51-ab84-904e88edbbbc-kube-api-access-kw6w9\") pod \"oauth-openshift-558db77b4-7wsdb\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.059128 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc9lq\" (UniqueName: \"kubernetes.io/projected/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-kube-api-access-rc9lq\") pod \"console-f9d7485db-b7fjd\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.070679 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzppv\" (UniqueName: \"kubernetes.io/projected/8de25024-e0c4-4a8c-b2c8-2435b8563c7b-kube-api-access-dzppv\") pod \"apiserver-7bbb656c7d-dfqhx\" (UID: \"8de25024-e0c4-4a8c-b2c8-2435b8563c7b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.087006 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p48dj\" (UniqueName: \"kubernetes.io/projected/0b0c22cf-0156-4cdb-a20f-1be43db0fedf-kube-api-access-p48dj\") pod \"ingress-operator-5b745b69d9-24hsx\" (UID: \"0b0c22cf-0156-4cdb-a20f-1be43db0fedf\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.108184 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk7qz\" (UniqueName: \"kubernetes.io/projected/96c3d581-243d-4d22-bf06-2771fec8192f-kube-api-access-kk7qz\") pod \"openshift-controller-manager-operator-756b6f6bc6-pgm2q\" (UID: \"96c3d581-243d-4d22-bf06-2771fec8192f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.129851 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tsq7\" (UniqueName: \"kubernetes.io/projected/8f806291-b5e1-4b3c-b010-e0f00881a445-kube-api-access-2tsq7\") pod \"machine-approver-56656f9798-h2pdm\" (UID: \"8f806291-b5e1-4b3c-b010-e0f00881a445\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.151468 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6545\" (UniqueName: \"kubernetes.io/projected/287782ba-f80b-4aae-a65a-9486f47a1e22-kube-api-access-w6545\") pod \"console-operator-58897d9998-wgkqf\" (UID: \"287782ba-f80b-4aae-a65a-9486f47a1e22\") " pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.152995 5035 request.go:700] Waited for 1.942818113s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/serviceaccounts/openshift-controller-manager-sa/token Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.170579 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdth5\" (UniqueName: \"kubernetes.io/projected/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-kube-api-access-xdth5\") pod \"controller-manager-879f6c89f-qb4qh\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.192230 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.194752 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2qvs\" (UniqueName: \"kubernetes.io/projected/368eaba6-c405-4ba6-afa9-13b32093851a-kube-api-access-w2qvs\") pod \"machine-api-operator-5694c8668f-9qdrb\" (UID: \"368eaba6-c405-4ba6-afa9-13b32093851a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.209772 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jqwf\" (UniqueName: \"kubernetes.io/projected/6a92f826-c0c7-4410-82bf-d2a2e0d3524d-kube-api-access-6jqwf\") pod \"etcd-operator-b45778765-fgjxt\" (UID: \"6a92f826-c0c7-4410-82bf-d2a2e0d3524d\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.226133 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.226386 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.233073 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvlhw\" (UniqueName: \"kubernetes.io/projected/f0096104-afd1-4e4f-80f7-8e45211965f1-kube-api-access-mvlhw\") pod \"openshift-config-operator-7777fb866f-5p9l9\" (UID: \"f0096104-afd1-4e4f-80f7-8e45211965f1\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.234185 5035 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.240183 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.247486 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-96thd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.254017 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.254046 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.268562 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.275326 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-m7jft"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.275463 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.275796 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.293376 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.298918 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.308794 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.316315 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.317539 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.328711 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.334454 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 09:30:06 crc kubenswrapper[5035]: W1002 09:30:06.338609 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac14f12c_aed8_4802_80dd_18a7aee2f254.slice/crio-1c693244c2c32bf2fc3e1ad5849489dcf6c6eb465a7a1e28921500d44c8ce6af WatchSource:0}: Error finding container 1c693244c2c32bf2fc3e1ad5849489dcf6c6eb465a7a1e28921500d44c8ce6af: Status 404 returned error can't find the container with id 1c693244c2c32bf2fc3e1ad5849489dcf6c6eb465a7a1e28921500d44c8ce6af Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.361054 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.374202 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.394297 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.415891 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.438947 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.447646 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1e1a7732-df59-4d74-9ab2-f57777fc8dfc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6br6r\" (UID: \"1e1a7732-df59-4d74-9ab2-f57777fc8dfc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.454662 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.460784 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf75h\" (UniqueName: \"kubernetes.io/projected/4fb71e4c-4c84-482f-819b-c32ba30776ec-kube-api-access-lf75h\") pod \"migrator-59844c95c7-dq99k\" (UID: \"4fb71e4c-4c84-482f-819b-c32ba30776ec\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.464233 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.464522 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.472966 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-pwdtp"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.493057 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd6wk\" (UniqueName: \"kubernetes.io/projected/e9af7278-65f4-471e-8b74-4305e9174038-kube-api-access-pd6wk\") pod \"collect-profiles-29323290-7w465\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.503130 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d4px\" (UniqueName: \"kubernetes.io/projected/3e222bb5-92b3-4a4d-833a-d383bdcc4515-kube-api-access-8d4px\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.504375 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-wgkqf"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.509612 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzt7g\" (UniqueName: \"kubernetes.io/projected/f1b15a92-86a9-4ab9-89d6-5de58811c55c-kube-api-access-rzt7g\") pod \"machine-config-controller-84d6567774-twhgt\" (UID: \"f1b15a92-86a9-4ab9-89d6-5de58811c55c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.537890 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25ng6\" (UniqueName: \"kubernetes.io/projected/86e0567f-4ddd-4034-8c0f-a9e7c009de8c-kube-api-access-25ng6\") pod \"multus-admission-controller-857f4d67dd-4ggtp\" (UID: \"86e0567f-4ddd-4034-8c0f-a9e7c009de8c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.548052 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndr5t\" (UniqueName: \"kubernetes.io/projected/977e3b39-af15-4940-ba0d-1e22466018c8-kube-api-access-ndr5t\") pod \"router-default-5444994796-rnzld\" (UID: \"977e3b39-af15-4940-ba0d-1e22466018c8\") " pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.582224 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc287200-ecc1-49ba-b81e-8b4a8c5aa96c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-9z8rz\" (UID: \"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.594377 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc4gd\" (UniqueName: \"kubernetes.io/projected/1717ec80-35f4-4622-8b21-abdf91627e98-kube-api-access-zc4gd\") pod \"ingress-canary-dcmtw\" (UID: \"1717ec80-35f4-4622-8b21-abdf91627e98\") " pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.619740 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dffdeca3-93e9-4773-bcb6-f3b63daab6ec-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-s46j8\" (UID: \"dffdeca3-93e9-4773-bcb6-f3b63daab6ec\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.646281 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.646745 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fgjxt"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.646287 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.651481 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.654609 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4r6p\" (UniqueName: \"kubernetes.io/projected/8336420c-cced-4dc7-bd94-e95ae46cdbc4-kube-api-access-m4r6p\") pod \"package-server-manager-789f6589d5-hwsxs\" (UID: \"8336420c-cced-4dc7-bd94-e95ae46cdbc4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.668641 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.679578 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.685478 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hphx\" (UniqueName: \"kubernetes.io/projected/10d07013-d746-4fc0-8485-64d95703dd12-kube-api-access-8hphx\") pod \"catalog-operator-68c6474976-ddmmt\" (UID: \"10d07013-d746-4fc0-8485-64d95703dd12\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.696198 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e222bb5-92b3-4a4d-833a-d383bdcc4515-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-ccztb\" (UID: \"3e222bb5-92b3-4a4d-833a-d383bdcc4515\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:06 crc kubenswrapper[5035]: W1002 09:30:06.704301 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a92f826_c0c7_4410_82bf_d2a2e0d3524d.slice/crio-b915ce735670de3fdd8e542fd7f1f48a1e9e6acc605c00036152c7648ecb7403 WatchSource:0}: Error finding container b915ce735670de3fdd8e542fd7f1f48a1e9e6acc605c00036152c7648ecb7403: Status 404 returned error can't find the container with id b915ce735670de3fdd8e542fd7f1f48a1e9e6acc605c00036152c7648ecb7403 Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.704985 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqjj5\" (UniqueName: \"kubernetes.io/projected/42daaa36-95f5-4fd6-bcbd-842b290ab83a-kube-api-access-jqjj5\") pod \"machine-config-operator-74547568cd-xcf7g\" (UID: \"42daaa36-95f5-4fd6-bcbd-842b290ab83a\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.732139 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.738136 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.771600 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-b7fjd"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.772448 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssh5v\" (UniqueName: \"kubernetes.io/projected/420d247b-10f8-4b8b-98dd-9fc29afdd60a-kube-api-access-ssh5v\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775105 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2j42\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-kube-api-access-r2j42\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775184 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/420d247b-10f8-4b8b-98dd-9fc29afdd60a-tmpfs\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775249 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775298 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc313ae8-012e-42ee-923e-29cfbb7ce593-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775324 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-srv-cert\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775398 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775442 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a7a7830-3a3d-4571-9dcd-ec0280e87e6a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-k2d9m\" (UID: \"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775485 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775550 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9stwd\" (UniqueName: \"kubernetes.io/projected/3f3cf046-2194-4792-aed0-7b08d65f6457-kube-api-access-9stwd\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775574 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5787d768-178b-4558-b2d8-8bc398a19c76-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775653 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775702 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxnxn\" (UniqueName: \"kubernetes.io/projected/5787d768-178b-4558-b2d8-8bc398a19c76-kube-api-access-nxnxn\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc313ae8-012e-42ee-923e-29cfbb7ce593-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775765 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-certificates\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775791 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5787d768-178b-4558-b2d8-8bc398a19c76-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775813 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h64r\" (UniqueName: \"kubernetes.io/projected/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-kube-api-access-8h64r\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775835 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/420d247b-10f8-4b8b-98dd-9fc29afdd60a-webhook-cert\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775910 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md4xn\" (UniqueName: \"kubernetes.io/projected/4a7a7830-3a3d-4571-9dcd-ec0280e87e6a-kube-api-access-md4xn\") pod \"control-plane-machine-set-operator-78cbb6b69f-k2d9m\" (UID: \"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775934 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-trusted-ca\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775955 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/420d247b-10f8-4b8b-98dd-9fc29afdd60a-apiservice-cert\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.775990 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-tls\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.776012 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-bound-sa-token\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.774267 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:06 crc kubenswrapper[5035]: E1002 09:30:06.778342 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.278328297 +0000 UTC m=+172.634672322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.783086 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-dcmtw" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.790427 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.802148 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.817706 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.821596 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" event={"ID":"287782ba-f80b-4aae-a65a-9486f47a1e22","Type":"ContainerStarted","Data":"7a7a2b75969efb7d0f6456a71a77820e9c5b6d4dbbc6edeaf660262dad0ccb1a"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.828663 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" event={"ID":"8de25024-e0c4-4a8c-b2c8-2435b8563c7b","Type":"ContainerStarted","Data":"ef7ee6ac647b43377c980799a1bb92d08e91779ef87caa41949f55d549c101be"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.844764 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-96thd"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.846259 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.846839 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.848733 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" event={"ID":"44ea8dc6-acd0-4b46-b90e-78428e973475","Type":"ContainerStarted","Data":"99f3afccb82e962d141dee8ed9f84a1f143a0a0164cc087ccffcc53d0a37d9c7"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.848771 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" event={"ID":"44ea8dc6-acd0-4b46-b90e-78428e973475","Type":"ContainerStarted","Data":"8e305d10fc02f5de3673a9495de6726a9ede34bc4e82fb35199947b018dc5ac2"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.860803 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" event={"ID":"6a92f826-c0c7-4410-82bf-d2a2e0d3524d","Type":"ContainerStarted","Data":"b915ce735670de3fdd8e542fd7f1f48a1e9e6acc605c00036152c7648ecb7403"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.868255 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-m7jft" event={"ID":"ac14f12c-aed8-4802-80dd-18a7aee2f254","Type":"ContainerStarted","Data":"bb4558747646137120d27766015975928c9ad1be7d5adf00e8891041b4ac5677"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.868290 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-m7jft" event={"ID":"ac14f12c-aed8-4802-80dd-18a7aee2f254","Type":"ContainerStarted","Data":"1c693244c2c32bf2fc3e1ad5849489dcf6c6eb465a7a1e28921500d44c8ce6af"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.868461 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-m7jft" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.869783 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-m7jft container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.869819 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m7jft" podUID="ac14f12c-aed8-4802-80dd-18a7aee2f254" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.875218 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-26669" event={"ID":"9975af5f-7bf6-41f9-919e-354f98ebc4d8","Type":"ContainerStarted","Data":"1bcd33d0d40a8fed2cdc11fc01fdf114a87a7ecbf1d8436feb5629212a16e425"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.875261 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-26669" event={"ID":"9975af5f-7bf6-41f9-919e-354f98ebc4d8","Type":"ContainerStarted","Data":"227733885c2b8c31d1fffc3ee48292aa70c264059f28d3c230958f4e8612a17b"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.876947 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877229 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssh5v\" (UniqueName: \"kubernetes.io/projected/420d247b-10f8-4b8b-98dd-9fc29afdd60a-kube-api-access-ssh5v\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877257 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-socket-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877294 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2j42\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-kube-api-access-r2j42\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877349 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-csi-data-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877365 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/420d247b-10f8-4b8b-98dd-9fc29afdd60a-tmpfs\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877445 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc313ae8-012e-42ee-923e-29cfbb7ce593-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877469 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-srv-cert\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877491 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1447ffd2-e487-4730-bc98-4fc864c0a4ca-signing-cabundle\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877550 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a90f3ffe-1cac-4b98-a7c0-526704066649-certs\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877576 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877593 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a7a7830-3a3d-4571-9dcd-ec0280e87e6a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-k2d9m\" (UID: \"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877612 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e276-6d06-4b2b-8972-40fc723ce9ef-config\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877634 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-mountpoint-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877647 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/88fc1d15-0b68-471a-9dc0-11b562d8deb0-metrics-tls\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877679 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a90f3ffe-1cac-4b98-a7c0-526704066649-node-bootstrap-token\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877747 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877789 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1447ffd2-e487-4730-bc98-4fc864c0a4ca-signing-key\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877804 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmrbz\" (UniqueName: \"kubernetes.io/projected/88fc1d15-0b68-471a-9dc0-11b562d8deb0-kube-api-access-xmrbz\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877822 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9stwd\" (UniqueName: \"kubernetes.io/projected/3f3cf046-2194-4792-aed0-7b08d65f6457-kube-api-access-9stwd\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877837 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bmvx\" (UniqueName: \"kubernetes.io/projected/40594fe7-b836-46fd-855f-7c3400356a24-kube-api-access-2bmvx\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877871 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5787d768-178b-4558-b2d8-8bc398a19c76-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877910 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f6e276-6d06-4b2b-8972-40fc723ce9ef-serving-cert\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.877931 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88fc1d15-0b68-471a-9dc0-11b562d8deb0-config-volume\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.878117 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.879768 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgl2h\" (UniqueName: \"kubernetes.io/projected/1447ffd2-e487-4730-bc98-4fc864c0a4ca-kube-api-access-qgl2h\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.880127 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxnxn\" (UniqueName: \"kubernetes.io/projected/5787d768-178b-4558-b2d8-8bc398a19c76-kube-api-access-nxnxn\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.880201 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc313ae8-012e-42ee-923e-29cfbb7ce593-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.882082 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.882690 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc313ae8-012e-42ee-923e-29cfbb7ce593-ca-trust-extracted\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.882934 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-certificates\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.882969 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5787d768-178b-4558-b2d8-8bc398a19c76-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.883000 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h64r\" (UniqueName: \"kubernetes.io/projected/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-kube-api-access-8h64r\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.883027 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/420d247b-10f8-4b8b-98dd-9fc29afdd60a-webhook-cert\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: E1002 09:30:06.883125 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.383105411 +0000 UTC m=+172.739449506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.883274 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" event={"ID":"8f806291-b5e1-4b3c-b010-e0f00881a445","Type":"ContainerStarted","Data":"5f0801abd623b32f0b62f1afcbe1fca1f6efbdd92316400c969c60f97b37618c"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.884113 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-plugins-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.884274 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/420d247b-10f8-4b8b-98dd-9fc29afdd60a-tmpfs\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.884844 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d26z6\" (UniqueName: \"kubernetes.io/projected/a90f3ffe-1cac-4b98-a7c0-526704066649-kube-api-access-d26z6\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.884899 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md4xn\" (UniqueName: \"kubernetes.io/projected/4a7a7830-3a3d-4571-9dcd-ec0280e87e6a-kube-api-access-md4xn\") pod \"control-plane-machine-set-operator-78cbb6b69f-k2d9m\" (UID: \"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885083 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-trusted-ca\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885111 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/420d247b-10f8-4b8b-98dd-9fc29afdd60a-apiservice-cert\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885147 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bl78\" (UniqueName: \"kubernetes.io/projected/48f6e276-6d06-4b2b-8972-40fc723ce9ef-kube-api-access-6bl78\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885181 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-tls\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885202 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-bound-sa-token\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885234 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885451 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc313ae8-012e-42ee-923e-29cfbb7ce593-installation-pull-secrets\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.885554 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-registration-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.886179 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-certificates\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.886324 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5787d768-178b-4558-b2d8-8bc398a19c76-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.886361 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-trusted-ca\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.893071 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-srv-cert\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.893626 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/420d247b-10f8-4b8b-98dd-9fc29afdd60a-webhook-cert\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.893678 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" event={"ID":"dec272a1-4b07-4d2d-99c5-776871d801ee","Type":"ContainerStarted","Data":"a924ac3d5a07e6750002c25a9d95a08333e3113dd5aea519ccccac77afa5a7ac"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.893760 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5787d768-178b-4558-b2d8-8bc398a19c76-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.895469 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/420d247b-10f8-4b8b-98dd-9fc29afdd60a-apiservice-cert\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.898747 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.899517 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-tls\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.903097 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" event={"ID":"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca","Type":"ContainerStarted","Data":"41ab8384a9ebec9ef819c42be0154a708adf217642ef8a0d90c27426365cf79d"} Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.914821 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4a7a7830-3a3d-4571-9dcd-ec0280e87e6a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-k2d9m\" (UID: \"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.914930 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2j42\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-kube-api-access-r2j42\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.916801 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7wsdb"] Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.961060 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.971789 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxnxn\" (UniqueName: \"kubernetes.io/projected/5787d768-178b-4558-b2d8-8bc398a19c76-kube-api-access-nxnxn\") pod \"kube-storage-version-migrator-operator-b67b599dd-k9ch7\" (UID: \"5787d768-178b-4558-b2d8-8bc398a19c76\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.983554 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx"] Oct 02 09:30:06 crc kubenswrapper[5035]: W1002 09:30:06.984694 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96c3d581_243d_4d22_bf06_2771fec8192f.slice/crio-e02db8398ed58c78cd9d7030b2c939dd8df630e2cd93c0ef17fa3ebdd08ea004 WatchSource:0}: Error finding container e02db8398ed58c78cd9d7030b2c939dd8df630e2cd93c0ef17fa3ebdd08ea004: Status 404 returned error can't find the container with id e02db8398ed58c78cd9d7030b2c939dd8df630e2cd93c0ef17fa3ebdd08ea004 Oct 02 09:30:06 crc kubenswrapper[5035]: W1002 09:30:06.985498 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5943cfb7_f43b_4a5f_8ac2_b39e8e898c56.slice/crio-4c75b3c68ef2d15d673d7f13dfaa4e38a661e9acff198faa869bb9fc79ff4524 WatchSource:0}: Error finding container 4c75b3c68ef2d15d673d7f13dfaa4e38a661e9acff198faa869bb9fc79ff4524: Status 404 returned error can't find the container with id 4c75b3c68ef2d15d673d7f13dfaa4e38a661e9acff198faa869bb9fc79ff4524 Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988235 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-plugins-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988274 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d26z6\" (UniqueName: \"kubernetes.io/projected/a90f3ffe-1cac-4b98-a7c0-526704066649-kube-api-access-d26z6\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988305 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bl78\" (UniqueName: \"kubernetes.io/projected/48f6e276-6d06-4b2b-8972-40fc723ce9ef-kube-api-access-6bl78\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988329 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-registration-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988350 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-socket-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988378 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-csi-data-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988408 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988427 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1447ffd2-e487-4730-bc98-4fc864c0a4ca-signing-cabundle\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988462 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a90f3ffe-1cac-4b98-a7c0-526704066649-certs\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988485 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e276-6d06-4b2b-8972-40fc723ce9ef-config\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988501 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-mountpoint-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/88fc1d15-0b68-471a-9dc0-11b562d8deb0-metrics-tls\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988542 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a90f3ffe-1cac-4b98-a7c0-526704066649-node-bootstrap-token\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988570 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1447ffd2-e487-4730-bc98-4fc864c0a4ca-signing-key\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988584 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmrbz\" (UniqueName: \"kubernetes.io/projected/88fc1d15-0b68-471a-9dc0-11b562d8deb0-kube-api-access-xmrbz\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988614 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bmvx\" (UniqueName: \"kubernetes.io/projected/40594fe7-b836-46fd-855f-7c3400356a24-kube-api-access-2bmvx\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988633 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f6e276-6d06-4b2b-8972-40fc723ce9ef-serving-cert\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988645 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88fc1d15-0b68-471a-9dc0-11b562d8deb0-config-volume\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.988692 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgl2h\" (UniqueName: \"kubernetes.io/projected/1447ffd2-e487-4730-bc98-4fc864c0a4ca-kube-api-access-qgl2h\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.989941 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-csi-data-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: E1002 09:30:06.990430 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.490415727 +0000 UTC m=+172.846759752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.991180 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-registration-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.991243 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-socket-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.991592 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/88fc1d15-0b68-471a-9dc0-11b562d8deb0-config-volume\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.991949 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48f6e276-6d06-4b2b-8972-40fc723ce9ef-config\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.992029 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-plugins-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.992120 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1447ffd2-e487-4730-bc98-4fc864c0a4ca-signing-cabundle\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.992128 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/40594fe7-b836-46fd-855f-7c3400356a24-mountpoint-dir\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.997111 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/88fc1d15-0b68-471a-9dc0-11b562d8deb0-metrics-tls\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.997246 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48f6e276-6d06-4b2b-8972-40fc723ce9ef-serving-cert\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.997796 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md4xn\" (UniqueName: \"kubernetes.io/projected/4a7a7830-3a3d-4571-9dcd-ec0280e87e6a-kube-api-access-md4xn\") pod \"control-plane-machine-set-operator-78cbb6b69f-k2d9m\" (UID: \"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:06 crc kubenswrapper[5035]: I1002 09:30:06.998264 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1447ffd2-e487-4730-bc98-4fc864c0a4ca-signing-key\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.001931 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9stwd\" (UniqueName: \"kubernetes.io/projected/3f3cf046-2194-4792-aed0-7b08d65f6457-kube-api-access-9stwd\") pod \"marketplace-operator-79b997595-jbph8\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.003649 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a90f3ffe-1cac-4b98-a7c0-526704066649-node-bootstrap-token\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.009644 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a90f3ffe-1cac-4b98-a7c0-526704066649-certs\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.047146 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.048624 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssh5v\" (UniqueName: \"kubernetes.io/projected/420d247b-10f8-4b8b-98dd-9fc29afdd60a-kube-api-access-ssh5v\") pod \"packageserver-d55dfcdfc-fl97h\" (UID: \"420d247b-10f8-4b8b-98dd-9fc29afdd60a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.049607 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-bound-sa-token\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.052210 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h64r\" (UniqueName: \"kubernetes.io/projected/37ff070a-2ff3-4ac5-9d23-f98f16895b1d-kube-api-access-8h64r\") pod \"olm-operator-6b444d44fb-5khtd\" (UID: \"37ff070a-2ff3-4ac5-9d23-f98f16895b1d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.068985 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.080818 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.088606 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9qdrb"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.089914 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.090269 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.590249567 +0000 UTC m=+172.946593592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.108598 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.109039 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgl2h\" (UniqueName: \"kubernetes.io/projected/1447ffd2-e487-4730-bc98-4fc864c0a4ca-kube-api-access-qgl2h\") pod \"service-ca-9c57cc56f-44nt7\" (UID: \"1447ffd2-e487-4730-bc98-4fc864c0a4ca\") " pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.115312 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bmvx\" (UniqueName: \"kubernetes.io/projected/40594fe7-b836-46fd-855f-7c3400356a24-kube-api-access-2bmvx\") pod \"csi-hostpathplugin-gpk2c\" (UID: \"40594fe7-b836-46fd-855f-7c3400356a24\") " pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.125600 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.133032 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmrbz\" (UniqueName: \"kubernetes.io/projected/88fc1d15-0b68-471a-9dc0-11b562d8deb0-kube-api-access-xmrbz\") pod \"dns-default-6cktt\" (UID: \"88fc1d15-0b68-471a-9dc0-11b562d8deb0\") " pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.138101 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.150356 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bl78\" (UniqueName: \"kubernetes.io/projected/48f6e276-6d06-4b2b-8972-40fc723ce9ef-kube-api-access-6bl78\") pod \"service-ca-operator-777779d784-xshqj\" (UID: \"48f6e276-6d06-4b2b-8972-40fc723ce9ef\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.155357 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.166860 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.178822 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d26z6\" (UniqueName: \"kubernetes.io/projected/a90f3ffe-1cac-4b98-a7c0-526704066649-kube-api-access-d26z6\") pod \"machine-config-server-6kjww\" (UID: \"a90f3ffe-1cac-4b98-a7c0-526704066649\") " pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.187971 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.190856 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.191222 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.69120665 +0000 UTC m=+173.047550675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.223373 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.225159 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-6kjww" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.291440 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.291609 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.791587527 +0000 UTC m=+173.147931552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.291771 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.292096 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.792086391 +0000 UTC m=+173.148430416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.344138 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.345953 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qb4qh"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.395916 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.396160 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.896145124 +0000 UTC m=+173.252489139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.497517 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.498036 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:07.998020024 +0000 UTC m=+173.354364049 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: W1002 09:30:07.515386 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda90f3ffe_1cac_4b98_a7c0_526704066649.slice/crio-61ba8d54c0edbfe16f0939450094e926ec06b71bde2c58d50a9456674c9d00c4 WatchSource:0}: Error finding container 61ba8d54c0edbfe16f0939450094e926ec06b71bde2c58d50a9456674c9d00c4: Status 404 returned error can't find the container with id 61ba8d54c0edbfe16f0939450094e926ec06b71bde2c58d50a9456674c9d00c4 Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.551827 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.551869 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.561969 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.590551 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.599085 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.599394 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.099376429 +0000 UTC m=+173.455720454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.710518 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.710925 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.210908738 +0000 UTC m=+173.567252753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.804279 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.813899 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.814418 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.314401855 +0000 UTC m=+173.670745880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.839752 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4ggtp"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.844166 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465"] Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.869733 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb"] Oct 02 09:30:07 crc kubenswrapper[5035]: W1002 09:30:07.870055 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42daaa36_95f5_4fd6_bcbd_842b290ab83a.slice/crio-e23909d5b0e09c302a4ae10febd04d62101974ffaf88ac4ff8812e4d5f89f6e6 WatchSource:0}: Error finding container e23909d5b0e09c302a4ae10febd04d62101974ffaf88ac4ff8812e4d5f89f6e6: Status 404 returned error can't find the container with id e23909d5b0e09c302a4ae10febd04d62101974ffaf88ac4ff8812e4d5f89f6e6 Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.915687 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.916047 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6kjww" event={"ID":"a90f3ffe-1cac-4b98-a7c0-526704066649","Type":"ContainerStarted","Data":"61ba8d54c0edbfe16f0939450094e926ec06b71bde2c58d50a9456674c9d00c4"} Oct 02 09:30:07 crc kubenswrapper[5035]: E1002 09:30:07.916301 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.416290865 +0000 UTC m=+173.772634880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.923627 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" event={"ID":"86e0567f-4ddd-4034-8c0f-a9e7c009de8c","Type":"ContainerStarted","Data":"e438608397eb9f9b4e0c0ae94545f524768a1cbdfd85d6434d31728bdd86e12e"} Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.926430 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" event={"ID":"8f806291-b5e1-4b3c-b010-e0f00881a445","Type":"ContainerStarted","Data":"e634c40ed2ec6b989c1a6d054cccd67b3fd33d3d16496936ab53d9d011cf0c9e"} Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.929497 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" event={"ID":"96c3d581-243d-4d22-bf06-2771fec8192f","Type":"ContainerStarted","Data":"e02db8398ed58c78cd9d7030b2c939dd8df630e2cd93c0ef17fa3ebdd08ea004"} Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.931373 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" event={"ID":"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c","Type":"ContainerStarted","Data":"20a6db96061649144f499764162089d1f77cd4fdd2fdb2d75b071dbcc1407609"} Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.935603 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" event={"ID":"287782ba-f80b-4aae-a65a-9486f47a1e22","Type":"ContainerStarted","Data":"7c68cf82a756722c90063c2c1bd92ebff8b64768b296d24ada4ba3b51500d687"} Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.936206 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.937857 5035 patch_prober.go:28] interesting pod/console-operator-58897d9998-wgkqf container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.937891 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" podUID="287782ba-f80b-4aae-a65a-9486f47a1e22" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.943058 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b7fjd" event={"ID":"b13e03a7-b033-441d-9f0b-2c6d85f3ab25","Type":"ContainerStarted","Data":"d0c1778af46655c95daeefaff5f8bc785a6a8f43d85a9b1e598c026786cb02b8"} Oct 02 09:30:07 crc kubenswrapper[5035]: I1002 09:30:07.950586 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" event={"ID":"e9af7278-65f4-471e-8b74-4305e9174038","Type":"ContainerStarted","Data":"e45b7fea71e431398e7b60dbe76b7d8e4a4bcb47b6c588b4d5270cd33838ef78"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.004286 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.004875 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" event={"ID":"dec272a1-4b07-4d2d-99c5-776871d801ee","Type":"ContainerStarted","Data":"397f35f33c56e1e165dce2537349b4af830d3ecf82a6235bce447679205329be"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.019228 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" event={"ID":"44ea8dc6-acd0-4b46-b90e-78428e973475","Type":"ContainerStarted","Data":"aa7ca91b58d482af0b0c5a55715071604949b8d6252a66b4f62c62281caf7493"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.024290 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.024653 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.524623611 +0000 UTC m=+173.880967636 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.024897 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.025159 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" event={"ID":"368eaba6-c405-4ba6-afa9-13b32093851a","Type":"ContainerStarted","Data":"5e7f9d8428204143cfb10244abe6b3dbb0fa578c7ee1163bd3bdd66602f148df"} Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.025291 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.52527924 +0000 UTC m=+173.881623265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: W1002 09:30:08.033558 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10d07013_d746_4fc0_8485_64d95703dd12.slice/crio-3d69c60a18c5125cbf00ea4ff7614995a2eb9d907bc1e335c04cf3cae9812244 WatchSource:0}: Error finding container 3d69c60a18c5125cbf00ea4ff7614995a2eb9d907bc1e335c04cf3cae9812244: Status 404 returned error can't find the container with id 3d69c60a18c5125cbf00ea4ff7614995a2eb9d907bc1e335c04cf3cae9812244 Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.049236 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" event={"ID":"dffdeca3-93e9-4773-bcb6-f3b63daab6ec","Type":"ContainerStarted","Data":"1fb6d70927bbd1cba1f5c096f4ca7c506b2d49acd56abd504b0dc0f67b7f1b20"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.052434 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.054412 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" event={"ID":"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae","Type":"ContainerStarted","Data":"725be3898fb7302ec7b03ab2a1f5b5f008b3a32b2caca5524da4ae9149ddaa96"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.055649 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.061356 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-44nt7"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.069635 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" event={"ID":"5ff717d1-2a4c-4f7f-a80b-c043a016d8ca","Type":"ContainerStarted","Data":"b6b323b8993a5ab265e26c154c1ad08e87f1e73ac6954d20fdbfebbe13174e49"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.086728 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6cktt"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.087328 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" event={"ID":"4fb71e4c-4c84-482f-819b-c32ba30776ec","Type":"ContainerStarted","Data":"b14456cddc86d675c0da1bb1948c15b22042ed119eca04ed5bfc0767eb08653c"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.088321 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" event={"ID":"8cabf8d0-11ee-4cb8-afb9-8dceb1461653","Type":"ContainerStarted","Data":"887ca1a3aa61f8ddee40e8e68692e215633820c4b20804b3f2af441ed396cb60"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.100114 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jbph8"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.100969 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-dcmtw"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.103024 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-xshqj"] Oct 02 09:30:08 crc kubenswrapper[5035]: W1002 09:30:08.120504 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8336420c_cced_4dc7_bd94_e95ae46cdbc4.slice/crio-c0d7a2e2597c7bb7b5d1d767b6978afe7ab372e56997f917bfc8fcc64e94bf99 WatchSource:0}: Error finding container c0d7a2e2597c7bb7b5d1d767b6978afe7ab372e56997f917bfc8fcc64e94bf99: Status 404 returned error can't find the container with id c0d7a2e2597c7bb7b5d1d767b6978afe7ab372e56997f917bfc8fcc64e94bf99 Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.127607 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.127681 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.627664725 +0000 UTC m=+173.984008750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.127946 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.128169 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.628162479 +0000 UTC m=+173.984506504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.148728 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-96thd" event={"ID":"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56","Type":"ContainerStarted","Data":"4c75b3c68ef2d15d673d7f13dfaa4e38a661e9acff198faa869bb9fc79ff4524"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.158433 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.162841 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" event={"ID":"1e1a7732-df59-4d74-9ab2-f57777fc8dfc","Type":"ContainerStarted","Data":"817b6053c4ec16045f381e851c5c68d9b37d2ecb45d80ecfba70432df38f1935"} Oct 02 09:30:08 crc kubenswrapper[5035]: W1002 09:30:08.186915 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f3cf046_2194_4792_aed0_7b08d65f6457.slice/crio-abd026cd2b3d67d9630a89502be01a1f8c8c06272323fa31ecf488569c85e2a3 WatchSource:0}: Error finding container abd026cd2b3d67d9630a89502be01a1f8c8c06272323fa31ecf488569c85e2a3: Status 404 returned error can't find the container with id abd026cd2b3d67d9630a89502be01a1f8c8c06272323fa31ecf488569c85e2a3 Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.191044 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" event={"ID":"f0096104-afd1-4e4f-80f7-8e45211965f1","Type":"ContainerStarted","Data":"d86712e89c3277755efafdd022c934908b5fc73b8198c3c1e52b420a7f8d7a61"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.191077 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gpk2c"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.191091 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.196032 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" event={"ID":"0b0c22cf-0156-4cdb-a20f-1be43db0fedf","Type":"ContainerStarted","Data":"757b1ae090d1889c5c60d7c58c794fb2fd644eb8fc87aad50496c0720dfb958b"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.197130 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" event={"ID":"42daaa36-95f5-4fd6-bcbd-842b290ab83a","Type":"ContainerStarted","Data":"e23909d5b0e09c302a4ae10febd04d62101974ffaf88ac4ff8812e4d5f89f6e6"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.197289 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h"] Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.204236 5035 generic.go:334] "Generic (PLEG): container finished" podID="8de25024-e0c4-4a8c-b2c8-2435b8563c7b" containerID="86b0986014e0052d2c5df294a3945373c4f23a80d18082b797f134fc35bfc294" exitCode=0 Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.204736 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" event={"ID":"8de25024-e0c4-4a8c-b2c8-2435b8563c7b","Type":"ContainerDied","Data":"86b0986014e0052d2c5df294a3945373c4f23a80d18082b797f134fc35bfc294"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.207719 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" event={"ID":"8abe751e-2b7d-4c51-ab84-904e88edbbbc","Type":"ContainerStarted","Data":"0ec7f1f1bf63ca5fdd985238a38421a19d274a4452944b97462a16c3575f58b2"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.211598 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-rnzld" event={"ID":"977e3b39-af15-4940-ba0d-1e22466018c8","Type":"ContainerStarted","Data":"245960df73c0c658c6b3a8d36e1d0f403fde494b7f9eb592b357fb19de8688c0"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.212858 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" event={"ID":"f1b15a92-86a9-4ab9-89d6-5de58811c55c","Type":"ContainerStarted","Data":"a6d7cc908bdc8174ead844cf2d850a7dfad934194a16c58c2e5f8387f23736a8"} Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.213388 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-m7jft container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.213422 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m7jft" podUID="ac14f12c-aed8-4802-80dd-18a7aee2f254" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 02 09:30:08 crc kubenswrapper[5035]: W1002 09:30:08.219994 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37ff070a_2ff3_4ac5_9d23_f98f16895b1d.slice/crio-6b68656895af33cbe750a679978e41b5856df9c3f1a701e4d9a69befd645d827 WatchSource:0}: Error finding container 6b68656895af33cbe750a679978e41b5856df9c3f1a701e4d9a69befd645d827: Status 404 returned error can't find the container with id 6b68656895af33cbe750a679978e41b5856df9c3f1a701e4d9a69befd645d827 Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.228723 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.230178 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.730156273 +0000 UTC m=+174.086500298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.329885 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.331148 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.831133266 +0000 UTC m=+174.187477291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.431209 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.431358 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.931337128 +0000 UTC m=+174.287681163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.431389 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.431780 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:08.93177146 +0000 UTC m=+174.288115485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.532171 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.532808 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.032781115 +0000 UTC m=+174.389125150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.634756 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.635197 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.13517605 +0000 UTC m=+174.491520135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.735685 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.735921 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.235902756 +0000 UTC m=+174.592246781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.742668 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.743936 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.743979 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.754640 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-pwdtp" podStartSLOduration=134.754624908 podStartE2EDuration="2m14.754624908s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:08.752289731 +0000 UTC m=+174.108633746" watchObservedRunningTime="2025-10-02 09:30:08.754624908 +0000 UTC m=+174.110968933" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.801240 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2nnft" podStartSLOduration=135.801223007 podStartE2EDuration="2m15.801223007s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:08.799668212 +0000 UTC m=+174.156012247" watchObservedRunningTime="2025-10-02 09:30:08.801223007 +0000 UTC m=+174.157567032" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.852093 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" podStartSLOduration=134.85206608 podStartE2EDuration="2m14.85206608s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:08.83272387 +0000 UTC m=+174.189067895" watchObservedRunningTime="2025-10-02 09:30:08.85206608 +0000 UTC m=+174.208410105" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.855129 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.856187 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.356140988 +0000 UTC m=+174.712485013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.929282 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-26669" podStartSLOduration=134.929268685 podStartE2EDuration="2m14.929268685s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:08.927141723 +0000 UTC m=+174.283485748" watchObservedRunningTime="2025-10-02 09:30:08.929268685 +0000 UTC m=+174.285612710" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.931054 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wxrhf" podStartSLOduration=134.931043976 podStartE2EDuration="2m14.931043976s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:08.888453063 +0000 UTC m=+174.244797088" watchObservedRunningTime="2025-10-02 09:30:08.931043976 +0000 UTC m=+174.287388001" Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.958220 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.958449 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.458406909 +0000 UTC m=+174.814750934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:08 crc kubenswrapper[5035]: I1002 09:30:08.958984 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:08 crc kubenswrapper[5035]: E1002 09:30:08.959360 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.459344956 +0000 UTC m=+174.815688971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.010165 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-m7jft" podStartSLOduration=135.010146847 podStartE2EDuration="2m15.010146847s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.009012474 +0000 UTC m=+174.365356499" watchObservedRunningTime="2025-10-02 09:30:09.010146847 +0000 UTC m=+174.366490872" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.012262 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-rnzld" podStartSLOduration=135.012253838 podStartE2EDuration="2m15.012253838s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:08.960908551 +0000 UTC m=+174.317252576" watchObservedRunningTime="2025-10-02 09:30:09.012253838 +0000 UTC m=+174.368597863" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.064675 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.064815 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.564793699 +0000 UTC m=+174.921137724 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.065146 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.065774 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.565732296 +0000 UTC m=+174.922076321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.166836 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.167808 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.667782691 +0000 UTC m=+175.024126716 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.174052 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.174488 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.674472085 +0000 UTC m=+175.030816110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.247676 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" event={"ID":"37ff070a-2ff3-4ac5-9d23-f98f16895b1d","Type":"ContainerStarted","Data":"78888f1ab64815fdb2cd62e8c242eb2af05274369ca03b673236aa7da76e2299"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.247724 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" event={"ID":"37ff070a-2ff3-4ac5-9d23-f98f16895b1d","Type":"ContainerStarted","Data":"6b68656895af33cbe750a679978e41b5856df9c3f1a701e4d9a69befd645d827"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.248472 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.253155 5035 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5khtd container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.253200 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" podUID="37ff070a-2ff3-4ac5-9d23-f98f16895b1d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.257604 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" event={"ID":"1e1a7732-df59-4d74-9ab2-f57777fc8dfc","Type":"ContainerStarted","Data":"b0e10c288f854235eb0cb4807d66168e749cade43aea0ffc61b5a1a264f46fb2"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.275370 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.275580 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.77551999 +0000 UTC m=+175.131864025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.275690 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.276170 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.776160659 +0000 UTC m=+175.132504694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.278830 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" event={"ID":"48f6e276-6d06-4b2b-8972-40fc723ce9ef","Type":"ContainerStarted","Data":"11dc49b9aa6b398335719b38a5d1cabbebe574db5a36803481adb9c83310ff3a"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.278873 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" event={"ID":"48f6e276-6d06-4b2b-8972-40fc723ce9ef","Type":"ContainerStarted","Data":"5ef5bfb594cc7e4e1299d8472fe27fa7cf8c37953f61fa4d5a1477547c85fafb"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.297521 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" event={"ID":"8336420c-cced-4dc7-bd94-e95ae46cdbc4","Type":"ContainerStarted","Data":"ed202c739f1007faadc3f5e6d822cfb59829ecaa0920ebc978ecd662f296a94f"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.297624 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" event={"ID":"8336420c-cced-4dc7-bd94-e95ae46cdbc4","Type":"ContainerStarted","Data":"c0d7a2e2597c7bb7b5d1d767b6978afe7ab372e56997f917bfc8fcc64e94bf99"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.308745 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" event={"ID":"10d07013-d746-4fc0-8485-64d95703dd12","Type":"ContainerStarted","Data":"cb32ef5d0769100795e4a8fe0335035cc23cd7c06d0d31c2290ed211a6b47e26"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.308796 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" event={"ID":"10d07013-d746-4fc0-8485-64d95703dd12","Type":"ContainerStarted","Data":"3d69c60a18c5125cbf00ea4ff7614995a2eb9d907bc1e335c04cf3cae9812244"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.310751 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.313913 5035 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-ddmmt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.313950 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" podUID="10d07013-d746-4fc0-8485-64d95703dd12" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.318181 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b7fjd" event={"ID":"b13e03a7-b033-441d-9f0b-2c6d85f3ab25","Type":"ContainerStarted","Data":"fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.331421 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" event={"ID":"368eaba6-c405-4ba6-afa9-13b32093851a","Type":"ContainerStarted","Data":"afa18770f68071844feabd9840768e26ae8c100552990f1880c922d19423d131"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.331458 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" event={"ID":"368eaba6-c405-4ba6-afa9-13b32093851a","Type":"ContainerStarted","Data":"e35a0fcfbc011b868119f4e060f02edcfa8c38b7bc454570d59758a84b9f39ef"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.356718 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" event={"ID":"3f3cf046-2194-4792-aed0-7b08d65f6457","Type":"ContainerStarted","Data":"12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.356767 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" event={"ID":"3f3cf046-2194-4792-aed0-7b08d65f6457","Type":"ContainerStarted","Data":"abd026cd2b3d67d9630a89502be01a1f8c8c06272323fa31ecf488569c85e2a3"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.357455 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.366724 5035 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jbph8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.366784 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.373923 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" event={"ID":"42daaa36-95f5-4fd6-bcbd-842b290ab83a","Type":"ContainerStarted","Data":"21a98143e8f85636108dfc6840bfe45f435881612deae5f10eda69b69d58318a"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.373959 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" event={"ID":"42daaa36-95f5-4fd6-bcbd-842b290ab83a","Type":"ContainerStarted","Data":"e28b1670d85605f27014ebe36120fa8a863a726a21971e05fb12d4164af8327d"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.377010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.378079 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.878059559 +0000 UTC m=+175.234403584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.381156 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" event={"ID":"8f806291-b5e1-4b3c-b010-e0f00881a445","Type":"ContainerStarted","Data":"13a99e1226659dfdac93c17b126be1c3047a3b3ac9e4a5ea6636bda8084bc8af"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.387121 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" event={"ID":"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae","Type":"ContainerStarted","Data":"0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.388099 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.391947 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-rnzld" event={"ID":"977e3b39-af15-4940-ba0d-1e22466018c8","Type":"ContainerStarted","Data":"4a1bcd80d17328f3df97de102497871bbdbf140b8dfec69c9c46ed7d90bb7bc0"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.393920 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" event={"ID":"420d247b-10f8-4b8b-98dd-9fc29afdd60a","Type":"ContainerStarted","Data":"331556b1562d7654ccd3a1f5a7609c5c87ef0997de8f09186364e6652293fe84"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.393954 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" event={"ID":"420d247b-10f8-4b8b-98dd-9fc29afdd60a","Type":"ContainerStarted","Data":"44c6e046452893ef0d9619dd80e4621e7870622ab8c03d9333ddb7f146550d98"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.394758 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.398030 5035 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-ldpj6 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.398070 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" podUID="da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.400192 5035 generic.go:334] "Generic (PLEG): container finished" podID="f0096104-afd1-4e4f-80f7-8e45211965f1" containerID="3e3edd874aca72c97c9418b0d6dda92b21c8c1aecf345c52134e03f2f7f758a4" exitCode=0 Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.400303 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" event={"ID":"f0096104-afd1-4e4f-80f7-8e45211965f1","Type":"ContainerDied","Data":"3e3edd874aca72c97c9418b0d6dda92b21c8c1aecf345c52134e03f2f7f758a4"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.402553 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" event={"ID":"3e222bb5-92b3-4a4d-833a-d383bdcc4515","Type":"ContainerStarted","Data":"7703ca8334360653585c38d8845b2d9bc6bf4a14a67788f880e33f3167bdac32"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.402585 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" event={"ID":"3e222bb5-92b3-4a4d-833a-d383bdcc4515","Type":"ContainerStarted","Data":"d631c666c5a2514b821e4279e08d1a7bb591027e13b9507a4ae97a3e3452612a"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.406670 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" event={"ID":"8abe751e-2b7d-4c51-ab84-904e88edbbbc","Type":"ContainerStarted","Data":"53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.406894 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.410162 5035 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-fl97h container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.410226 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" podUID="420d247b-10f8-4b8b-98dd-9fc29afdd60a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.410375 5035 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7wsdb container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.410430 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" podUID="8abe751e-2b7d-4c51-ab84-904e88edbbbc" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.418605 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-96thd" event={"ID":"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56","Type":"ContainerStarted","Data":"6c6708f623a41f5816a5a917f3f33eadf09c859c36ccaac1807f71afae4bfe5b"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.423390 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" event={"ID":"6a92f826-c0c7-4410-82bf-d2a2e0d3524d","Type":"ContainerStarted","Data":"d420d7ced5079adb4a0d668d1804d9702616fc27d8dbc9bd0d9f052a6518def2"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.438381 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" event={"ID":"86e0567f-4ddd-4034-8c0f-a9e7c009de8c","Type":"ContainerStarted","Data":"71eb46a64bd25309f383737ad2b1eade13e3bf729734b98f7097ff95d69af9c6"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.440487 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" event={"ID":"5787d768-178b-4558-b2d8-8bc398a19c76","Type":"ContainerStarted","Data":"4758959c8afdc6406afd49e86f1b01d7897e1b2dd79ae16a391c739050b5b484"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.440517 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" event={"ID":"5787d768-178b-4558-b2d8-8bc398a19c76","Type":"ContainerStarted","Data":"de106bf1ec31a6d8c7e80f84e404d043d6d00ae05068885e85255288751f09ec"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.458910 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" event={"ID":"40594fe7-b836-46fd-855f-7c3400356a24","Type":"ContainerStarted","Data":"10dc20010a263177db3473b8915b736b82dd3ee6e9c75528636e19867e3c2a1f"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.464200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" event={"ID":"96c3d581-243d-4d22-bf06-2771fec8192f","Type":"ContainerStarted","Data":"450bf351ac2067092c07f2a464908f6d5a7bbb036b10b6bb9d65ae70b51eb090"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.470898 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9qdrb" podStartSLOduration=135.470877667 podStartE2EDuration="2m15.470877667s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.467638613 +0000 UTC m=+174.823982638" watchObservedRunningTime="2025-10-02 09:30:09.470877667 +0000 UTC m=+174.827221692" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.471500 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" event={"ID":"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a","Type":"ContainerStarted","Data":"18d991f45c56d88d8a14efcd121eb0c09255140885180798a947395924f120a8"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.471552 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" event={"ID":"4a7a7830-3a3d-4571-9dcd-ec0280e87e6a","Type":"ContainerStarted","Data":"1cfc9cc475281ef3cb805c5ccc63661205a4d9530cd43933cc1e48fa80bd6368"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.476103 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-6kjww" event={"ID":"a90f3ffe-1cac-4b98-a7c0-526704066649","Type":"ContainerStarted","Data":"c010fc15f3687d12a16fe3c881e6a3eed1d028bd24436e4290990e0a548054a3"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.478264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.486670 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:09.986648883 +0000 UTC m=+175.342992908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.498764 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" event={"ID":"8de25024-e0c4-4a8c-b2c8-2435b8563c7b","Type":"ContainerStarted","Data":"161c6b54c5915fbffe50be1ac2dd88c4b91bc58dc94c2353b399ece2815efd3a"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.518605 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xcf7g" podStartSLOduration=135.518583498 podStartE2EDuration="2m15.518583498s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.514268113 +0000 UTC m=+174.870612138" watchObservedRunningTime="2025-10-02 09:30:09.518583498 +0000 UTC m=+174.874927533" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.523849 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" event={"ID":"1447ffd2-e487-4730-bc98-4fc864c0a4ca","Type":"ContainerStarted","Data":"a0350ac346efe53b2f282aaf72755688254857e1ab3835e517e5d0a27dee6e40"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.524101 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" event={"ID":"1447ffd2-e487-4730-bc98-4fc864c0a4ca","Type":"ContainerStarted","Data":"6ac19c8536ee5e3d184fc7543743be7ddc4d389ec6e3856ea776693c22ab734f"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.539385 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" event={"ID":"dffdeca3-93e9-4773-bcb6-f3b63daab6ec","Type":"ContainerStarted","Data":"338466433b559c5a6b9a64373a4df273bd22c18d8b04e0403a6960e97fdf209f"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.541474 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" event={"ID":"f1b15a92-86a9-4ab9-89d6-5de58811c55c","Type":"ContainerStarted","Data":"60b2bdb43d7c4deabea9cc7412d8bf5f740751cc412c60e3e964efbef899435a"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.541512 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" event={"ID":"f1b15a92-86a9-4ab9-89d6-5de58811c55c","Type":"ContainerStarted","Data":"fa33aaf5f6653f61f5f488d6b027216cecb9567c72269eb74beabf702fc9650a"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.557644 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" event={"ID":"e9af7278-65f4-471e-8b74-4305e9174038","Type":"ContainerStarted","Data":"c628f3e20bddacc8a7c57934cf125c44f4f12b1d512663a8434c00e6a5a13794"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.578448 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" event={"ID":"4fb71e4c-4c84-482f-819b-c32ba30776ec","Type":"ContainerStarted","Data":"18f6662a8c592e0395fdb1fd1977a3db9c603c70bb96aea8a92ce8002ddec066"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.578510 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" event={"ID":"4fb71e4c-4c84-482f-819b-c32ba30776ec","Type":"ContainerStarted","Data":"3664640107eb36b96dea6f6dcd88d83323fe1622aaa321016707226e623c72a1"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.579925 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.581396 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.081374956 +0000 UTC m=+175.437718991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.592106 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-xshqj" podStartSLOduration=134.592086826 podStartE2EDuration="2m14.592086826s" podCreationTimestamp="2025-10-02 09:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.558614747 +0000 UTC m=+174.914958772" watchObservedRunningTime="2025-10-02 09:30:09.592086826 +0000 UTC m=+174.948430851" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.593587 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-fgjxt" podStartSLOduration=135.59358062 podStartE2EDuration="2m15.59358062s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.58737457 +0000 UTC m=+174.943718595" watchObservedRunningTime="2025-10-02 09:30:09.59358062 +0000 UTC m=+174.949924645" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.631947 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dcmtw" event={"ID":"1717ec80-35f4-4622-8b21-abdf91627e98","Type":"ContainerStarted","Data":"80a9d0c5fe62d9df3b12d0bfa235b3d2e53b61115828c721df5d93e3ffc05242"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.632005 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-dcmtw" event={"ID":"1717ec80-35f4-4622-8b21-abdf91627e98","Type":"ContainerStarted","Data":"b2e3fa3c1ddcec5455390e62e8563549163411658a089ad6195eceb5cc2c51ad"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.641782 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" event={"ID":"0b0c22cf-0156-4cdb-a20f-1be43db0fedf","Type":"ContainerStarted","Data":"fcdcaba4e2d7424066c3ef4f9e4cfdd180e124fa9cf257c4b27889f39245f3bd"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.641837 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" event={"ID":"0b0c22cf-0156-4cdb-a20f-1be43db0fedf","Type":"ContainerStarted","Data":"ad185cc10b47436b6e6f7fad8a3e8bf6c4f792905d3b8765f4a70b89e1431654"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.671112 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6cktt" event={"ID":"88fc1d15-0b68-471a-9dc0-11b562d8deb0","Type":"ContainerStarted","Data":"c4ee773000ea1733dce8164e87b989d07574ba9984ff566375427da1e540e246"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.683024 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.685109 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.185091239 +0000 UTC m=+175.541435334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.693778 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" event={"ID":"bc287200-ecc1-49ba-b81e-8b4a8c5aa96c","Type":"ContainerStarted","Data":"1b70730a114cd18d5a79267ed5f2662d922b2771e95d8d2ac5c460e5d0ceee30"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.699409 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" podStartSLOduration=135.699388333 podStartE2EDuration="2m15.699388333s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.632241289 +0000 UTC m=+174.988585324" watchObservedRunningTime="2025-10-02 09:30:09.699388333 +0000 UTC m=+175.055732358" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.714642 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" event={"ID":"8cabf8d0-11ee-4cb8-afb9-8dceb1461653","Type":"ContainerStarted","Data":"7604f7d6bcf6ef92f6d526c19948e01af44108975e2a68dea6755a670d21a40e"} Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.714689 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.737141 5035 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qb4qh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.737195 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" podUID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.760228 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" podStartSLOduration=135.760206654 podStartE2EDuration="2m15.760206654s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.72274753 +0000 UTC m=+175.079091565" watchObservedRunningTime="2025-10-02 09:30:09.760206654 +0000 UTC m=+175.116550679" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.760356 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" podStartSLOduration=135.760350928 podStartE2EDuration="2m15.760350928s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.759205775 +0000 UTC m=+175.115549790" watchObservedRunningTime="2025-10-02 09:30:09.760350928 +0000 UTC m=+175.116694963" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.768975 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:09 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:09 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:09 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.769022 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.786030 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.787051 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.287035721 +0000 UTC m=+175.643379746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.838977 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" podStartSLOduration=134.838962835 podStartE2EDuration="2m14.838962835s" podCreationTimestamp="2025-10-02 09:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.837453291 +0000 UTC m=+175.193797316" watchObservedRunningTime="2025-10-02 09:30:09.838962835 +0000 UTC m=+175.195306860" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.839956 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h2pdm" podStartSLOduration=136.839947583 podStartE2EDuration="2m16.839947583s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.785566799 +0000 UTC m=+175.141910824" watchObservedRunningTime="2025-10-02 09:30:09.839947583 +0000 UTC m=+175.196291608" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.887484 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.890796 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.390781995 +0000 UTC m=+175.747126020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.904278 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.904604 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.927288 5035 patch_prober.go:28] interesting pod/apiserver-76f77b778f-26669 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]log ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]etcd ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/generic-apiserver-start-informers ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/max-in-flight-filter ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 02 09:30:09 crc kubenswrapper[5035]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 02 09:30:09 crc kubenswrapper[5035]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/project.openshift.io-projectcache ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/openshift.io-startinformers ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 02 09:30:09 crc kubenswrapper[5035]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 02 09:30:09 crc kubenswrapper[5035]: livez check failed Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.927338 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-26669" podUID="9975af5f-7bf6-41f9-919e-354f98ebc4d8" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.930734 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" podStartSLOduration=135.930716391 podStartE2EDuration="2m15.930716391s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.928674592 +0000 UTC m=+175.285018617" watchObservedRunningTime="2025-10-02 09:30:09.930716391 +0000 UTC m=+175.287060416" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.931729 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" podStartSLOduration=135.93172168 podStartE2EDuration="2m15.93172168s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.901242158 +0000 UTC m=+175.257586183" watchObservedRunningTime="2025-10-02 09:30:09.93172168 +0000 UTC m=+175.288065705" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.944612 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6br6r" podStartSLOduration=135.944597283 podStartE2EDuration="2m15.944597283s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.944280844 +0000 UTC m=+175.300624869" watchObservedRunningTime="2025-10-02 09:30:09.944597283 +0000 UTC m=+175.300941308" Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.989011 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.989245 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.489198695 +0000 UTC m=+175.845542720 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.989488 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:09 crc kubenswrapper[5035]: E1002 09:30:09.990620 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.490606645 +0000 UTC m=+175.846950670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:09 crc kubenswrapper[5035]: I1002 09:30:09.991864 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-b7fjd" podStartSLOduration=135.991846761 podStartE2EDuration="2m15.991846761s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:09.991646455 +0000 UTC m=+175.347990480" watchObservedRunningTime="2025-10-02 09:30:09.991846761 +0000 UTC m=+175.348190786" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.010974 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-wgkqf" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.031005 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-ccztb" podStartSLOduration=136.030989755 podStartE2EDuration="2m16.030989755s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.028125082 +0000 UTC m=+175.384469117" watchObservedRunningTime="2025-10-02 09:30:10.030989755 +0000 UTC m=+175.387333780" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.077267 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" podStartSLOduration=136.077248414 podStartE2EDuration="2m16.077248414s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.075142803 +0000 UTC m=+175.431486828" watchObservedRunningTime="2025-10-02 09:30:10.077248414 +0000 UTC m=+175.433592439" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.091403 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.091790 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.591774255 +0000 UTC m=+175.948118280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.103022 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-dcmtw" podStartSLOduration=7.10300472 podStartE2EDuration="7.10300472s" podCreationTimestamp="2025-10-02 09:30:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.102491315 +0000 UTC m=+175.458835340" watchObservedRunningTime="2025-10-02 09:30:10.10300472 +0000 UTC m=+175.459348745" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.193195 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.193583 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.693566592 +0000 UTC m=+176.049910667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.199638 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-dq99k" podStartSLOduration=136.199581146 podStartE2EDuration="2m16.199581146s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.152876224 +0000 UTC m=+175.509220259" watchObservedRunningTime="2025-10-02 09:30:10.199581146 +0000 UTC m=+175.555925171" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.238576 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k9ch7" podStartSLOduration=136.238558505 podStartE2EDuration="2m16.238558505s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.201879853 +0000 UTC m=+175.558223878" watchObservedRunningTime="2025-10-02 09:30:10.238558505 +0000 UTC m=+175.594902530" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.295165 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.295453 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.795437911 +0000 UTC m=+176.151781936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.336493 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-twhgt" podStartSLOduration=136.33647264 podStartE2EDuration="2m16.33647264s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.294380091 +0000 UTC m=+175.650724116" watchObservedRunningTime="2025-10-02 09:30:10.33647264 +0000 UTC m=+175.692816665" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.366487 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-k2d9m" podStartSLOduration=136.366472578 podStartE2EDuration="2m16.366472578s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.335481841 +0000 UTC m=+175.691825866" watchObservedRunningTime="2025-10-02 09:30:10.366472578 +0000 UTC m=+175.722816603" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.368062 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-44nt7" podStartSLOduration=135.368056874 podStartE2EDuration="2m15.368056874s" podCreationTimestamp="2025-10-02 09:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.365795689 +0000 UTC m=+175.722139714" watchObservedRunningTime="2025-10-02 09:30:10.368056874 +0000 UTC m=+175.724400899" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.389208 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-6kjww" podStartSLOduration=6.389191806 podStartE2EDuration="6.389191806s" podCreationTimestamp="2025-10-02 09:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.386505978 +0000 UTC m=+175.742850003" watchObservedRunningTime="2025-10-02 09:30:10.389191806 +0000 UTC m=+175.745535831" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.396757 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.397091 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.897073974 +0000 UTC m=+176.253417999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.433422 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-9z8rz" podStartSLOduration=136.433403526 podStartE2EDuration="2m16.433403526s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.429756011 +0000 UTC m=+175.786100056" watchObservedRunningTime="2025-10-02 09:30:10.433403526 +0000 UTC m=+175.789747541" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.480193 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pgm2q" podStartSLOduration=136.480177691 podStartE2EDuration="2m16.480177691s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.47877253 +0000 UTC m=+175.835116555" watchObservedRunningTime="2025-10-02 09:30:10.480177691 +0000 UTC m=+175.836521716" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.498766 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.498912 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.998887082 +0000 UTC m=+176.355231107 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.499047 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.499442 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:10.999428038 +0000 UTC m=+176.355772073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.523373 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" podStartSLOduration=10.523357791 podStartE2EDuration="10.523357791s" podCreationTimestamp="2025-10-02 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.52194967 +0000 UTC m=+175.878293695" watchObservedRunningTime="2025-10-02 09:30:10.523357791 +0000 UTC m=+175.879701816" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.549855 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-s46j8" podStartSLOduration=136.549835136 podStartE2EDuration="2m16.549835136s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.546905443 +0000 UTC m=+175.903249468" watchObservedRunningTime="2025-10-02 09:30:10.549835136 +0000 UTC m=+175.906179171" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.588854 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-24hsx" podStartSLOduration=136.588839776 podStartE2EDuration="2m16.588839776s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.588491126 +0000 UTC m=+175.944835161" watchObservedRunningTime="2025-10-02 09:30:10.588839776 +0000 UTC m=+175.945183791" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.599906 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.600220 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.100206095 +0000 UTC m=+176.456550120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.647182 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" podStartSLOduration=136.647161274 podStartE2EDuration="2m16.647161274s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.64700177 +0000 UTC m=+176.003345795" watchObservedRunningTime="2025-10-02 09:30:10.647161274 +0000 UTC m=+176.003505299" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.702047 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.702525 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.202504367 +0000 UTC m=+176.558848392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.750808 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:10 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:10 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:10 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.750865 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.762421 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6cktt" event={"ID":"88fc1d15-0b68-471a-9dc0-11b562d8deb0","Type":"ContainerStarted","Data":"b5de36bbef924fdf81187e1bb5e98543b29bf86b315952aac16f38985d8969a4"} Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.762490 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6cktt" event={"ID":"88fc1d15-0b68-471a-9dc0-11b562d8deb0","Type":"ContainerStarted","Data":"eb92b94140c04aa42c2e3984aa8d219a889b30ff5b264db9752727aea82d9065"} Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.763326 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.764994 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" event={"ID":"86e0567f-4ddd-4034-8c0f-a9e7c009de8c","Type":"ContainerStarted","Data":"549d3e3d46f701869bc318e835751fa912cc81fb12aeb721b0893744a20169fc"} Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.784955 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" event={"ID":"8336420c-cced-4dc7-bd94-e95ae46cdbc4","Type":"ContainerStarted","Data":"905fd625b2a5ee5532f1a7af4f668a989695f167719f533c424a0bc8337c90c3"} Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.785850 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.798798 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-96thd" event={"ID":"5943cfb7-f43b-4a5f-8ac2-b39e8e898c56","Type":"ContainerStarted","Data":"e1f12ac8b2f364b0e98568892131a25a195f42ff2f1f5218fbffe903cc3976f0"} Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.801818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" event={"ID":"40594fe7-b836-46fd-855f-7c3400356a24","Type":"ContainerStarted","Data":"4556b2ccc3f068ac865180f45247bd996b01087e6cb7aef79ae1f2b2e24efb37"} Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.802933 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.803105 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.303083609 +0000 UTC m=+176.659427644 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.803365 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.803727 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.303717207 +0000 UTC m=+176.660061232 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.808884 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" event={"ID":"f0096104-afd1-4e4f-80f7-8e45211965f1","Type":"ContainerStarted","Data":"dea6b8c001f11d15ee8353180e3f0707ccfc0911312e6577beb2c0171c26a9bb"} Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.808925 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.810776 5035 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-fl97h container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.810820 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" podUID="420d247b-10f8-4b8b-98dd-9fc29afdd60a" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.811000 5035 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jbph8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.811120 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.811264 5035 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5khtd container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.811361 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" podUID="37ff070a-2ff3-4ac5-9d23-f98f16895b1d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.811776 5035 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qb4qh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.811813 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" podUID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.834723 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.859550 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.873464 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ddmmt" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.903027 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6cktt" podStartSLOduration=6.903010962 podStartE2EDuration="6.903010962s" podCreationTimestamp="2025-10-02 09:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.847791334 +0000 UTC m=+176.204135369" watchObservedRunningTime="2025-10-02 09:30:10.903010962 +0000 UTC m=+176.259354987" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.905849 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:10 crc kubenswrapper[5035]: E1002 09:30:10.906880 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.406861084 +0000 UTC m=+176.763205109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.985501 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-4ggtp" podStartSLOduration=136.9854828 podStartE2EDuration="2m16.9854828s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.903153656 +0000 UTC m=+176.259497681" watchObservedRunningTime="2025-10-02 09:30:10.9854828 +0000 UTC m=+176.341826825" Oct 02 09:30:10 crc kubenswrapper[5035]: I1002 09:30:10.987366 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" podStartSLOduration=136.987351044 podStartE2EDuration="2m16.987351044s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:10.977598002 +0000 UTC m=+176.333942027" watchObservedRunningTime="2025-10-02 09:30:10.987351044 +0000 UTC m=+176.343695069" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.009486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.009903 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.509890677 +0000 UTC m=+176.866234692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.110807 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.111234 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.611216641 +0000 UTC m=+176.967560666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.150438 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-96thd" podStartSLOduration=137.150418076 podStartE2EDuration="2m17.150418076s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:11.061935034 +0000 UTC m=+176.418279059" watchObservedRunningTime="2025-10-02 09:30:11.150418076 +0000 UTC m=+176.506762111" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.152587 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" podStartSLOduration=137.152575068 podStartE2EDuration="2m17.152575068s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:11.146156582 +0000 UTC m=+176.502500607" watchObservedRunningTime="2025-10-02 09:30:11.152575068 +0000 UTC m=+176.508919093" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.194650 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.195004 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.214184 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.214563 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.714525922 +0000 UTC m=+177.070869947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.315169 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.315563 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.815549537 +0000 UTC m=+177.171893562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.417217 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.417512 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:11.917501419 +0000 UTC m=+177.273845444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.517887 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.518216 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.018201535 +0000 UTC m=+177.374545560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.618765 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.619062 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.119050305 +0000 UTC m=+177.475394330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.719912 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.720284 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.220269906 +0000 UTC m=+177.576613931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.747995 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:11 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:11 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:11 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.748042 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.813483 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" event={"ID":"40594fe7-b836-46fd-855f-7c3400356a24","Type":"ContainerStarted","Data":"dfbfb746ea2bcc4614b3e77f0cb83f618ca183cf13acea29031229689a23d964"} Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.817467 5035 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-jbph8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.817516 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.821709 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.821963 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.32194983 +0000 UTC m=+177.678293855 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.825773 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-fl97h" Oct 02 09:30:11 crc kubenswrapper[5035]: I1002 09:30:11.923307 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:11 crc kubenswrapper[5035]: E1002 09:30:11.924787 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.424765177 +0000 UTC m=+177.781109212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.025462 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.025836 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.525823623 +0000 UTC m=+177.882167648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.032443 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.127064 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.127394 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.627380113 +0000 UTC m=+177.983724138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.228475 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.228765 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.728754048 +0000 UTC m=+178.085098073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.234027 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-khfkq"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.234863 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.237132 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.256973 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khfkq"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.329591 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.329733 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.829708141 +0000 UTC m=+178.186052166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.330525 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc4zg\" (UniqueName: \"kubernetes.io/projected/af1e80e7-afaf-4fa5-993d-54fb39e36285-kube-api-access-xc4zg\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.330688 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-catalog-content\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.330835 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-utilities\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.330991 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.331270 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.831261106 +0000 UTC m=+178.187605131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.431811 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.431948 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.931930071 +0000 UTC m=+178.288274096 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.432109 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc4zg\" (UniqueName: \"kubernetes.io/projected/af1e80e7-afaf-4fa5-993d-54fb39e36285-kube-api-access-xc4zg\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.432154 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-catalog-content\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.432210 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-utilities\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.432596 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-catalog-content\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.432825 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-utilities\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.432853 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.433000 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:12.932991702 +0000 UTC m=+178.289335727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.437470 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nbpds"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.445832 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.450595 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.458469 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbpds"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.509796 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc4zg\" (UniqueName: \"kubernetes.io/projected/af1e80e7-afaf-4fa5-993d-54fb39e36285-kube-api-access-xc4zg\") pod \"certified-operators-khfkq\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.516246 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5p9l9" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.533448 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.533810 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-utilities\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.534004 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnd4q\" (UniqueName: \"kubernetes.io/projected/dbb7f782-87ae-4f6c-8642-520a483497c0-kube-api-access-rnd4q\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.534085 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-catalog-content\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.534306 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:13.034286355 +0000 UTC m=+178.390630380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.550331 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.597362 5035 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.622771 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tqtkq"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.623721 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.636252 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.636514 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-utilities\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.636721 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnd4q\" (UniqueName: \"kubernetes.io/projected/dbb7f782-87ae-4f6c-8642-520a483497c0-kube-api-access-rnd4q\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.637148 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 09:30:13.137132723 +0000 UTC m=+178.493476758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-2rf7w" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.637735 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-catalog-content\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.638156 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-catalog-content\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.638199 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqtkq"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.637591 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-utilities\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.684424 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnd4q\" (UniqueName: \"kubernetes.io/projected/dbb7f782-87ae-4f6c-8642-520a483497c0-kube-api-access-rnd4q\") pod \"community-operators-nbpds\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.727744 5035 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T09:30:12.597486815Z","Handler":null,"Name":""} Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.739175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.739349 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-utilities\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.739380 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65mg2\" (UniqueName: \"kubernetes.io/projected/1b6bb9c6-cf32-45b3-8031-f8c787649c42-kube-api-access-65mg2\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.739423 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-catalog-content\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: E1002 09:30:12.739518 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 09:30:13.239502787 +0000 UTC m=+178.595846812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.744672 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:12 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:12 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:12 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.744734 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.759993 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.772706 5035 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.772756 5035 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.829745 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sxgqx"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.830671 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.849318 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-catalog-content\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.851143 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.851205 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-utilities\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.851245 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65mg2\" (UniqueName: \"kubernetes.io/projected/1b6bb9c6-cf32-45b3-8031-f8c787649c42-kube-api-access-65mg2\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.850397 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-catalog-content\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.852010 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-utilities\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.853136 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sxgqx"] Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.884477 5035 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.884515 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.904307 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65mg2\" (UniqueName: \"kubernetes.io/projected/1b6bb9c6-cf32-45b3-8031-f8c787649c42-kube-api-access-65mg2\") pod \"certified-operators-tqtkq\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.907859 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" event={"ID":"40594fe7-b836-46fd-855f-7c3400356a24","Type":"ContainerStarted","Data":"28fcb11fbdfb64e3590c8c0ea0e60247637ae05bcc92b4383ba7186bf627818d"} Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.907888 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" event={"ID":"40594fe7-b836-46fd-855f-7c3400356a24","Type":"ContainerStarted","Data":"7a98ea11cdca8de0a533de7c8820843d2a3e3dd83b35eac4c00c71c91952e0ee"} Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.928709 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-dfqhx" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.934004 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-2rf7w\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.937266 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-gpk2c" podStartSLOduration=9.937250563 podStartE2EDuration="9.937250563s" podCreationTimestamp="2025-10-02 09:30:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:12.936052898 +0000 UTC m=+178.292396923" watchObservedRunningTime="2025-10-02 09:30:12.937250563 +0000 UTC m=+178.293594588" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.953608 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-utilities\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.953654 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-catalog-content\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.953714 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd5nv\" (UniqueName: \"kubernetes.io/projected/67c99626-0b6b-4255-80f9-836f764375c4-kube-api-access-fd5nv\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.953861 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:30:12 crc kubenswrapper[5035]: I1002 09:30:12.977488 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.056799 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.057161 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-utilities\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.057228 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-catalog-content\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.057409 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd5nv\" (UniqueName: \"kubernetes.io/projected/67c99626-0b6b-4255-80f9-836f764375c4-kube-api-access-fd5nv\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.061399 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-catalog-content\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.061825 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-utilities\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.113243 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd5nv\" (UniqueName: \"kubernetes.io/projected/67c99626-0b6b-4255-80f9-836f764375c4-kube-api-access-fd5nv\") pod \"community-operators-sxgqx\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.113377 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-khfkq"] Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.136402 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.230022 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.261169 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbpds"] Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.484679 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tqtkq"] Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.586380 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.587509 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.590269 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.590994 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.591176 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.622139 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2rf7w"] Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.683055 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sxgqx"] Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.743367 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:13 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:13 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:13 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.743415 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.769157 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.769214 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.870982 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.871052 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.871141 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.908934 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.917034 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.927592 5035 generic.go:334] "Generic (PLEG): container finished" podID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerID="a495158c37215164ea59f085a639c7609ccd9abbf1928574837db3ee27ed12a5" exitCode=0 Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.927865 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqtkq" event={"ID":"1b6bb9c6-cf32-45b3-8031-f8c787649c42","Type":"ContainerDied","Data":"a495158c37215164ea59f085a639c7609ccd9abbf1928574837db3ee27ed12a5"} Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.927909 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqtkq" event={"ID":"1b6bb9c6-cf32-45b3-8031-f8c787649c42","Type":"ContainerStarted","Data":"27635ce7066902204fcbdfafc862d894a5b7efefbe0e77496186d52ce83dfd16"} Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.931096 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" event={"ID":"cc313ae8-012e-42ee-923e-29cfbb7ce593","Type":"ContainerStarted","Data":"ad71f63a35ee209b57ef42a0eab79889e198b5d3da438a59b9b108bdb2875563"} Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.931404 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.948228 5035 generic.go:334] "Generic (PLEG): container finished" podID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerID="17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db" exitCode=0 Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.948336 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbpds" event={"ID":"dbb7f782-87ae-4f6c-8642-520a483497c0","Type":"ContainerDied","Data":"17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db"} Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.948368 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbpds" event={"ID":"dbb7f782-87ae-4f6c-8642-520a483497c0","Type":"ContainerStarted","Data":"fa1cbc8c64084d4bc5f358bf9704936ac82420ca3ac331508df7740a5184d359"} Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.964265 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxgqx" event={"ID":"67c99626-0b6b-4255-80f9-836f764375c4","Type":"ContainerStarted","Data":"ae0db716a12060d4a59f25da5401fe6851228a90425227ae59612fbde85b0af5"} Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.971741 5035 generic.go:334] "Generic (PLEG): container finished" podID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerID="6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9" exitCode=0 Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.972587 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfkq" event={"ID":"af1e80e7-afaf-4fa5-993d-54fb39e36285","Type":"ContainerDied","Data":"6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9"} Oct 02 09:30:13 crc kubenswrapper[5035]: I1002 09:30:13.972616 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfkq" event={"ID":"af1e80e7-afaf-4fa5-993d-54fb39e36285","Type":"ContainerStarted","Data":"11cae9c1df010b2b35a856efd55eb86cf4b82ba53eb627edb7e49a1608fff1da"} Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.170402 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.199183 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 09:30:14 crc kubenswrapper[5035]: W1002 09:30:14.199455 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podeb7123e7_7420_4d3c_b6ad_c52dcb00c2e6.slice/crio-cd0fd3fe60bf3759587eb927056c71cca4c4d9570d3ceec8a6ffcb20a7308f70 WatchSource:0}: Error finding container cd0fd3fe60bf3759587eb927056c71cca4c4d9570d3ceec8a6ffcb20a7308f70: Status 404 returned error can't find the container with id cd0fd3fe60bf3759587eb927056c71cca4c4d9570d3ceec8a6ffcb20a7308f70 Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.230037 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jg5zd"] Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.230991 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.233363 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.250145 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg5zd"] Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.380472 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-catalog-content\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.380744 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj6nx\" (UniqueName: \"kubernetes.io/projected/d727546b-168e-4cd6-9ffc-ddd2623cf743-kube-api-access-dj6nx\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.380789 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-utilities\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.482254 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj6nx\" (UniqueName: \"kubernetes.io/projected/d727546b-168e-4cd6-9ffc-ddd2623cf743-kube-api-access-dj6nx\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.482340 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-utilities\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.482478 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-catalog-content\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.483174 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-utilities\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.483392 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-catalog-content\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.502723 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj6nx\" (UniqueName: \"kubernetes.io/projected/d727546b-168e-4cd6-9ffc-ddd2623cf743-kube-api-access-dj6nx\") pod \"redhat-marketplace-jg5zd\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.543690 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.647130 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-47v4z"] Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.648903 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.653567 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47v4z"] Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.742476 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:14 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:14 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:14 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.742862 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.766865 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg5zd"] Oct 02 09:30:14 crc kubenswrapper[5035]: W1002 09:30:14.775478 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd727546b_168e_4cd6_9ffc_ddd2623cf743.slice/crio-56f420064518690a20000b1bdfdbceee0d3fe1a2eb281dd6bda40edf1f5cf8ef WatchSource:0}: Error finding container 56f420064518690a20000b1bdfdbceee0d3fe1a2eb281dd6bda40edf1f5cf8ef: Status 404 returned error can't find the container with id 56f420064518690a20000b1bdfdbceee0d3fe1a2eb281dd6bda40edf1f5cf8ef Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.787993 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9222\" (UniqueName: \"kubernetes.io/projected/09530660-adcc-4a75-8550-116b440990d5-kube-api-access-f9222\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.788057 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-catalog-content\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.788077 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-utilities\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.890339 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-catalog-content\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.890393 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-utilities\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.890466 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9222\" (UniqueName: \"kubernetes.io/projected/09530660-adcc-4a75-8550-116b440990d5-kube-api-access-f9222\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.891456 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-utilities\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.897320 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-catalog-content\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.913836 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9222\" (UniqueName: \"kubernetes.io/projected/09530660-adcc-4a75-8550-116b440990d5-kube-api-access-f9222\") pod \"redhat-marketplace-47v4z\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.922633 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.927956 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-26669" Oct 02 09:30:14 crc kubenswrapper[5035]: I1002 09:30:14.975127 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.065671 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" event={"ID":"cc313ae8-012e-42ee-923e-29cfbb7ce593","Type":"ContainerStarted","Data":"c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765"} Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.066588 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.090688 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" podStartSLOduration=141.090671952 podStartE2EDuration="2m21.090671952s" podCreationTimestamp="2025-10-02 09:27:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:15.090025824 +0000 UTC m=+180.446369849" watchObservedRunningTime="2025-10-02 09:30:15.090671952 +0000 UTC m=+180.447015977" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.105877 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6","Type":"ContainerStarted","Data":"419c02b17edd4b11aeeebf1396ee810ec4d7fe29469aa231c22667fde9427bb2"} Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.105944 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6","Type":"ContainerStarted","Data":"cd0fd3fe60bf3759587eb927056c71cca4c4d9570d3ceec8a6ffcb20a7308f70"} Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.121818 5035 generic.go:334] "Generic (PLEG): container finished" podID="67c99626-0b6b-4255-80f9-836f764375c4" containerID="0b24df4157d616cb2ac9c2ddaa6becf68eda381e7edfacf07f0735ce34036447" exitCode=0 Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.121899 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxgqx" event={"ID":"67c99626-0b6b-4255-80f9-836f764375c4","Type":"ContainerDied","Data":"0b24df4157d616cb2ac9c2ddaa6becf68eda381e7edfacf07f0735ce34036447"} Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.126729 5035 generic.go:334] "Generic (PLEG): container finished" podID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerID="153e61c0cb7b0fc50dc5000e828df335bd6ae03d9d6023e8d47734534cc37edd" exitCode=0 Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.128399 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg5zd" event={"ID":"d727546b-168e-4cd6-9ffc-ddd2623cf743","Type":"ContainerDied","Data":"153e61c0cb7b0fc50dc5000e828df335bd6ae03d9d6023e8d47734534cc37edd"} Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.128428 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg5zd" event={"ID":"d727546b-168e-4cd6-9ffc-ddd2623cf743","Type":"ContainerStarted","Data":"56f420064518690a20000b1bdfdbceee0d3fe1a2eb281dd6bda40edf1f5cf8ef"} Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.145729 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.145713946 podStartE2EDuration="2.145713946s" podCreationTimestamp="2025-10-02 09:30:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:15.129663611 +0000 UTC m=+180.486007636" watchObservedRunningTime="2025-10-02 09:30:15.145713946 +0000 UTC m=+180.502057971" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.359394 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47v4z"] Oct 02 09:30:15 crc kubenswrapper[5035]: W1002 09:30:15.397660 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09530660_adcc_4a75_8550_116b440990d5.slice/crio-c3d239ff5b23503971c28fefc885dae36a385edbb792b98334854ad7e1ab570a WatchSource:0}: Error finding container c3d239ff5b23503971c28fefc885dae36a385edbb792b98334854ad7e1ab570a: Status 404 returned error can't find the container with id c3d239ff5b23503971c28fefc885dae36a385edbb792b98334854ad7e1ab570a Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.432279 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g8hgc"] Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.433234 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.438552 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.448806 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8hgc"] Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.511914 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-utilities\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.512000 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvzpb\" (UniqueName: \"kubernetes.io/projected/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-kube-api-access-cvzpb\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.512027 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-catalog-content\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.613957 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-utilities\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.614001 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.614031 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvzpb\" (UniqueName: \"kubernetes.io/projected/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-kube-api-access-cvzpb\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.614050 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-catalog-content\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.614518 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-catalog-content\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.614941 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-utilities\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.620063 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/452af00b-602d-43ab-a345-5453d6aebcf0-metrics-certs\") pod \"network-metrics-daemon-hzsjk\" (UID: \"452af00b-602d-43ab-a345-5453d6aebcf0\") " pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.650996 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvzpb\" (UniqueName: \"kubernetes.io/projected/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-kube-api-access-cvzpb\") pod \"redhat-operators-g8hgc\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.742523 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:15 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:15 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:15 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.742593 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.752567 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.823044 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mx6nb"] Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.828343 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.832631 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mx6nb"] Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.887837 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hzsjk" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.917805 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-catalog-content\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.917874 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-utilities\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.917893 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7lhj\" (UniqueName: \"kubernetes.io/projected/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-kube-api-access-x7lhj\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.985772 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-m7jft container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.985830 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m7jft" podUID="ac14f12c-aed8-4802-80dd-18a7aee2f254" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.986215 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-m7jft container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 02 09:30:15 crc kubenswrapper[5035]: I1002 09:30:15.986267 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-m7jft" podUID="ac14f12c-aed8-4802-80dd-18a7aee2f254" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.019470 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-catalog-content\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.019588 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-utilities\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.019614 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7lhj\" (UniqueName: \"kubernetes.io/projected/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-kube-api-access-x7lhj\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.020569 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-catalog-content\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.022105 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-utilities\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.042412 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7lhj\" (UniqueName: \"kubernetes.io/projected/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-kube-api-access-x7lhj\") pod \"redhat-operators-mx6nb\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.135346 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g8hgc"] Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.195193 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.220511 5035 generic.go:334] "Generic (PLEG): container finished" podID="eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6" containerID="419c02b17edd4b11aeeebf1396ee810ec4d7fe29469aa231c22667fde9427bb2" exitCode=0 Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.229681 5035 generic.go:334] "Generic (PLEG): container finished" podID="09530660-adcc-4a75-8550-116b440990d5" containerID="d2a80cfccc6444c4d96de46e6607c43c0901fc2311d844a1de059eaee8aff6c0" exitCode=0 Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.235918 5035 generic.go:334] "Generic (PLEG): container finished" podID="e9af7278-65f4-471e-8b74-4305e9174038" containerID="c628f3e20bddacc8a7c57934cf125c44f4f12b1d512663a8434c00e6a5a13794" exitCode=0 Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.242885 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.242917 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.242931 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hzsjk"] Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.242947 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6","Type":"ContainerDied","Data":"419c02b17edd4b11aeeebf1396ee810ec4d7fe29469aa231c22667fde9427bb2"} Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.242968 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47v4z" event={"ID":"09530660-adcc-4a75-8550-116b440990d5","Type":"ContainerDied","Data":"d2a80cfccc6444c4d96de46e6607c43c0901fc2311d844a1de059eaee8aff6c0"} Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.242967 5035 patch_prober.go:28] interesting pod/console-f9d7485db-b7fjd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.243013 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-b7fjd" podUID="b13e03a7-b033-441d-9f0b-2c6d85f3ab25" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.242982 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47v4z" event={"ID":"09530660-adcc-4a75-8550-116b440990d5","Type":"ContainerStarted","Data":"c3d239ff5b23503971c28fefc885dae36a385edbb792b98334854ad7e1ab570a"} Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.243265 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" event={"ID":"e9af7278-65f4-471e-8b74-4305e9174038","Type":"ContainerDied","Data":"c628f3e20bddacc8a7c57934cf125c44f4f12b1d512663a8434c00e6a5a13794"} Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.444350 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.545596 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mx6nb"] Oct 02 09:30:16 crc kubenswrapper[5035]: W1002 09:30:16.558595 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda3fd37ad_729b_475d_8ff1_74e2d4a0b19b.slice/crio-e78dc2c5c410dfeff72cd478b72d8433a0bc82c29ddfb5f2e5cf22ed75fd6aeb WatchSource:0}: Error finding container e78dc2c5c410dfeff72cd478b72d8433a0bc82c29ddfb5f2e5cf22ed75fd6aeb: Status 404 returned error can't find the container with id e78dc2c5c410dfeff72cd478b72d8433a0bc82c29ddfb5f2e5cf22ed75fd6aeb Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.697751 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.698764 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.712820 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.713021 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.716943 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.739056 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.745555 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:16 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:16 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:16 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.745601 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.846018 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f1801415-437b-4db2-9a7d-dccf260a0184-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.846118 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f1801415-437b-4db2-9a7d-dccf260a0184-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.948000 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f1801415-437b-4db2-9a7d-dccf260a0184-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.948095 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f1801415-437b-4db2-9a7d-dccf260a0184-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.948116 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f1801415-437b-4db2-9a7d-dccf260a0184-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:16 crc kubenswrapper[5035]: I1002 09:30:16.968483 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f1801415-437b-4db2-9a7d-dccf260a0184-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.058894 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.116719 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.148290 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5khtd" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.250439 5035 generic.go:334] "Generic (PLEG): container finished" podID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerID="dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e" exitCode=0 Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.250507 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8hgc" event={"ID":"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e","Type":"ContainerDied","Data":"dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e"} Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.250880 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8hgc" event={"ID":"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e","Type":"ContainerStarted","Data":"34045aec80839f15e3a5bea86eba72f20ca19123e4fb6d18c3515260fcafd91a"} Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.259190 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" event={"ID":"452af00b-602d-43ab-a345-5453d6aebcf0","Type":"ContainerStarted","Data":"f0116771977304c81ca4b453910b6c9e63d3c5f389b33ef34cc8b42e23e27847"} Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.259231 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" event={"ID":"452af00b-602d-43ab-a345-5453d6aebcf0","Type":"ContainerStarted","Data":"7dd452f5cc17e7e92e3a7f0ca98c5641b6892c7535245f51ab43e9a532f88f8b"} Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.274886 5035 generic.go:334] "Generic (PLEG): container finished" podID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerID="8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b" exitCode=0 Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.276070 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mx6nb" event={"ID":"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b","Type":"ContainerDied","Data":"8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b"} Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.276131 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mx6nb" event={"ID":"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b","Type":"ContainerStarted","Data":"e78dc2c5c410dfeff72cd478b72d8433a0bc82c29ddfb5f2e5cf22ed75fd6aeb"} Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.392812 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 09:30:17 crc kubenswrapper[5035]: W1002 09:30:17.417615 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podf1801415_437b_4db2_9a7d_dccf260a0184.slice/crio-5a07f801062625f7cc62203549b25aedcfb73de8befdb3c1b9a1ce07c0bb2e95 WatchSource:0}: Error finding container 5a07f801062625f7cc62203549b25aedcfb73de8befdb3c1b9a1ce07c0bb2e95: Status 404 returned error can't find the container with id 5a07f801062625f7cc62203549b25aedcfb73de8befdb3c1b9a1ce07c0bb2e95 Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.579391 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.653438 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.667803 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd6wk\" (UniqueName: \"kubernetes.io/projected/e9af7278-65f4-471e-8b74-4305e9174038-kube-api-access-pd6wk\") pod \"e9af7278-65f4-471e-8b74-4305e9174038\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.667957 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9af7278-65f4-471e-8b74-4305e9174038-secret-volume\") pod \"e9af7278-65f4-471e-8b74-4305e9174038\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.668004 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9af7278-65f4-471e-8b74-4305e9174038-config-volume\") pod \"e9af7278-65f4-471e-8b74-4305e9174038\" (UID: \"e9af7278-65f4-471e-8b74-4305e9174038\") " Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.669256 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9af7278-65f4-471e-8b74-4305e9174038-config-volume" (OuterVolumeSpecName: "config-volume") pod "e9af7278-65f4-471e-8b74-4305e9174038" (UID: "e9af7278-65f4-471e-8b74-4305e9174038"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.676976 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9af7278-65f4-471e-8b74-4305e9174038-kube-api-access-pd6wk" (OuterVolumeSpecName: "kube-api-access-pd6wk") pod "e9af7278-65f4-471e-8b74-4305e9174038" (UID: "e9af7278-65f4-471e-8b74-4305e9174038"). InnerVolumeSpecName "kube-api-access-pd6wk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.683470 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9af7278-65f4-471e-8b74-4305e9174038-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e9af7278-65f4-471e-8b74-4305e9174038" (UID: "e9af7278-65f4-471e-8b74-4305e9174038"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.746065 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:17 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:17 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:17 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.746132 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.773306 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kube-api-access\") pod \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.773376 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kubelet-dir\") pod \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\" (UID: \"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6\") " Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.773676 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9af7278-65f4-471e-8b74-4305e9174038-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.773712 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9af7278-65f4-471e-8b74-4305e9174038-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.773726 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd6wk\" (UniqueName: \"kubernetes.io/projected/e9af7278-65f4-471e-8b74-4305e9174038-kube-api-access-pd6wk\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.773787 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6" (UID: "eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.785953 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6" (UID: "eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.875201 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:17 crc kubenswrapper[5035]: I1002 09:30:17.875229 5035 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.292123 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hzsjk" event={"ID":"452af00b-602d-43ab-a345-5453d6aebcf0","Type":"ContainerStarted","Data":"2759a434d5e73583fbb30290044c12bcfe79635b6307af5730c8c51951ad390c"} Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.295788 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" event={"ID":"e9af7278-65f4-471e-8b74-4305e9174038","Type":"ContainerDied","Data":"e45b7fea71e431398e7b60dbe76b7d8e4a4bcb47b6c588b4d5270cd33838ef78"} Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.295833 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e45b7fea71e431398e7b60dbe76b7d8e4a4bcb47b6c588b4d5270cd33838ef78" Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.295841 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465" Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.306393 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6","Type":"ContainerDied","Data":"cd0fd3fe60bf3759587eb927056c71cca4c4d9570d3ceec8a6ffcb20a7308f70"} Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.306432 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd0fd3fe60bf3759587eb927056c71cca4c4d9570d3ceec8a6ffcb20a7308f70" Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.306459 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.308441 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hzsjk" podStartSLOduration=145.308419969 podStartE2EDuration="2m25.308419969s" podCreationTimestamp="2025-10-02 09:27:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:18.308188303 +0000 UTC m=+183.664532328" watchObservedRunningTime="2025-10-02 09:30:18.308419969 +0000 UTC m=+183.664763994" Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.308716 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f1801415-437b-4db2-9a7d-dccf260a0184","Type":"ContainerStarted","Data":"5a07f801062625f7cc62203549b25aedcfb73de8befdb3c1b9a1ce07c0bb2e95"} Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.742558 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:18 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:18 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:18 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:18 crc kubenswrapper[5035]: I1002 09:30:18.742627 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:19 crc kubenswrapper[5035]: I1002 09:30:19.227036 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6cktt" Oct 02 09:30:19 crc kubenswrapper[5035]: I1002 09:30:19.335373 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f1801415-437b-4db2-9a7d-dccf260a0184","Type":"ContainerStarted","Data":"f02144cbb7724aec9cb09d44a0bc4afaa36e542793df0e63ca43427ce2219210"} Oct 02 09:30:19 crc kubenswrapper[5035]: I1002 09:30:19.354999 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.354983792 podStartE2EDuration="3.354983792s" podCreationTimestamp="2025-10-02 09:30:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:30:19.348678269 +0000 UTC m=+184.705022314" watchObservedRunningTime="2025-10-02 09:30:19.354983792 +0000 UTC m=+184.711327817" Oct 02 09:30:19 crc kubenswrapper[5035]: I1002 09:30:19.741744 5035 patch_prober.go:28] interesting pod/router-default-5444994796-rnzld container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 09:30:19 crc kubenswrapper[5035]: [-]has-synced failed: reason withheld Oct 02 09:30:19 crc kubenswrapper[5035]: [+]process-running ok Oct 02 09:30:19 crc kubenswrapper[5035]: healthz check failed Oct 02 09:30:19 crc kubenswrapper[5035]: I1002 09:30:19.741806 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-rnzld" podUID="977e3b39-af15-4940-ba0d-1e22466018c8" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 09:30:20 crc kubenswrapper[5035]: I1002 09:30:20.346110 5035 generic.go:334] "Generic (PLEG): container finished" podID="f1801415-437b-4db2-9a7d-dccf260a0184" containerID="f02144cbb7724aec9cb09d44a0bc4afaa36e542793df0e63ca43427ce2219210" exitCode=0 Oct 02 09:30:20 crc kubenswrapper[5035]: I1002 09:30:20.346160 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f1801415-437b-4db2-9a7d-dccf260a0184","Type":"ContainerDied","Data":"f02144cbb7724aec9cb09d44a0bc4afaa36e542793df0e63ca43427ce2219210"} Oct 02 09:30:20 crc kubenswrapper[5035]: I1002 09:30:20.743348 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:20 crc kubenswrapper[5035]: I1002 09:30:20.745595 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-rnzld" Oct 02 09:30:25 crc kubenswrapper[5035]: I1002 09:30:25.537796 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:30:25 crc kubenswrapper[5035]: I1002 09:30:25.538347 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:30:25 crc kubenswrapper[5035]: I1002 09:30:25.986116 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-m7jft container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 02 09:30:25 crc kubenswrapper[5035]: I1002 09:30:25.986495 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-m7jft" podUID="ac14f12c-aed8-4802-80dd-18a7aee2f254" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 02 09:30:25 crc kubenswrapper[5035]: I1002 09:30:25.986116 5035 patch_prober.go:28] interesting pod/downloads-7954f5f757-m7jft container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" start-of-body= Oct 02 09:30:25 crc kubenswrapper[5035]: I1002 09:30:25.986575 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-m7jft" podUID="ac14f12c-aed8-4802-80dd-18a7aee2f254" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.15:8080/\": dial tcp 10.217.0.15:8080: connect: connection refused" Oct 02 09:30:26 crc kubenswrapper[5035]: I1002 09:30:26.245525 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:26 crc kubenswrapper[5035]: I1002 09:30:26.249234 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:30:27 crc kubenswrapper[5035]: I1002 09:30:27.688023 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:27 crc kubenswrapper[5035]: I1002 09:30:27.847893 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f1801415-437b-4db2-9a7d-dccf260a0184-kubelet-dir\") pod \"f1801415-437b-4db2-9a7d-dccf260a0184\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " Oct 02 09:30:27 crc kubenswrapper[5035]: I1002 09:30:27.847992 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1801415-437b-4db2-9a7d-dccf260a0184-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f1801415-437b-4db2-9a7d-dccf260a0184" (UID: "f1801415-437b-4db2-9a7d-dccf260a0184"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:30:27 crc kubenswrapper[5035]: I1002 09:30:27.848014 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f1801415-437b-4db2-9a7d-dccf260a0184-kube-api-access\") pod \"f1801415-437b-4db2-9a7d-dccf260a0184\" (UID: \"f1801415-437b-4db2-9a7d-dccf260a0184\") " Oct 02 09:30:27 crc kubenswrapper[5035]: I1002 09:30:27.848431 5035 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f1801415-437b-4db2-9a7d-dccf260a0184-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:27 crc kubenswrapper[5035]: I1002 09:30:27.855217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1801415-437b-4db2-9a7d-dccf260a0184-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f1801415-437b-4db2-9a7d-dccf260a0184" (UID: "f1801415-437b-4db2-9a7d-dccf260a0184"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:30:27 crc kubenswrapper[5035]: I1002 09:30:27.950883 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f1801415-437b-4db2-9a7d-dccf260a0184-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 09:30:28 crc kubenswrapper[5035]: I1002 09:30:28.398228 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"f1801415-437b-4db2-9a7d-dccf260a0184","Type":"ContainerDied","Data":"5a07f801062625f7cc62203549b25aedcfb73de8befdb3c1b9a1ce07c0bb2e95"} Oct 02 09:30:28 crc kubenswrapper[5035]: I1002 09:30:28.398482 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a07f801062625f7cc62203549b25aedcfb73de8befdb3c1b9a1ce07c0bb2e95" Oct 02 09:30:28 crc kubenswrapper[5035]: I1002 09:30:28.398292 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 09:30:32 crc kubenswrapper[5035]: I1002 09:30:32.984307 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:30:35 crc kubenswrapper[5035]: I1002 09:30:35.994055 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-m7jft" Oct 02 09:30:42 crc kubenswrapper[5035]: E1002 09:30:42.850045 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 09:30:42 crc kubenswrapper[5035]: E1002 09:30:42.850550 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-65mg2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tqtkq_openshift-marketplace(1b6bb9c6-cf32-45b3-8031-f8c787649c42): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:42 crc kubenswrapper[5035]: E1002 09:30:42.852339 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tqtkq" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" Oct 02 09:30:43 crc kubenswrapper[5035]: I1002 09:30:43.509387 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 09:30:46 crc kubenswrapper[5035]: E1002 09:30:46.350264 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tqtkq" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" Oct 02 09:30:46 crc kubenswrapper[5035]: I1002 09:30:46.797234 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hwsxs" Oct 02 09:30:47 crc kubenswrapper[5035]: E1002 09:30:47.271998 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 09:30:47 crc kubenswrapper[5035]: E1002 09:30:47.272182 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x7lhj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mx6nb_openshift-marketplace(a3fd37ad-729b-475d-8ff1-74e2d4a0b19b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:47 crc kubenswrapper[5035]: E1002 09:30:47.273379 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mx6nb" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.119094 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mx6nb" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.209975 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.210461 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rnd4q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nbpds_openshift-marketplace(dbb7f782-87ae-4f6c-8642-520a483497c0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.211958 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nbpds" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.224122 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.224264 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xc4zg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-khfkq_openshift-marketplace(af1e80e7-afaf-4fa5-993d-54fb39e36285): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.225575 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-khfkq" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.248713 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.248893 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fd5nv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-sxgqx_openshift-marketplace(67c99626-0b6b-4255-80f9-836f764375c4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.250268 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-sxgqx" podUID="67c99626-0b6b-4255-80f9-836f764375c4" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.527990 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.528181 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cvzpb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-g8hgc_openshift-marketplace(6d6a45a6-8f75-47ba-86d9-cb2ff35c288e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:48 crc kubenswrapper[5035]: E1002 09:30:48.529481 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-g8hgc" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.158203 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-khfkq" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.158393 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-g8hgc" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.158468 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-sxgqx" podUID="67c99626-0b6b-4255-80f9-836f764375c4" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.158555 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nbpds" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.805485 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.805929 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dj6nx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jg5zd_openshift-marketplace(d727546b-168e-4cd6-9ffc-ddd2623cf743): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.807155 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jg5zd" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.816601 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.816811 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f9222,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-47v4z_openshift-marketplace(09530660-adcc-4a75-8550-116b440990d5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 09:30:51 crc kubenswrapper[5035]: E1002 09:30:51.818070 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-47v4z" podUID="09530660-adcc-4a75-8550-116b440990d5" Oct 02 09:30:52 crc kubenswrapper[5035]: E1002 09:30:52.565772 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-47v4z" podUID="09530660-adcc-4a75-8550-116b440990d5" Oct 02 09:30:52 crc kubenswrapper[5035]: E1002 09:30:52.567216 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jg5zd" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" Oct 02 09:30:55 crc kubenswrapper[5035]: I1002 09:30:55.537987 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:30:55 crc kubenswrapper[5035]: I1002 09:30:55.538069 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:31:00 crc kubenswrapper[5035]: I1002 09:31:00.604824 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mx6nb" event={"ID":"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b","Type":"ContainerStarted","Data":"0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae"} Oct 02 09:31:01 crc kubenswrapper[5035]: I1002 09:31:01.612457 5035 generic.go:334] "Generic (PLEG): container finished" podID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerID="0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae" exitCode=0 Oct 02 09:31:01 crc kubenswrapper[5035]: I1002 09:31:01.612561 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mx6nb" event={"ID":"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b","Type":"ContainerDied","Data":"0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae"} Oct 02 09:31:01 crc kubenswrapper[5035]: I1002 09:31:01.616127 5035 generic.go:334] "Generic (PLEG): container finished" podID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerID="3e60c6de61ceab13fb59c89f9079b7b1a86d2fa3766742c11a60190d50d680ae" exitCode=0 Oct 02 09:31:01 crc kubenswrapper[5035]: I1002 09:31:01.616177 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqtkq" event={"ID":"1b6bb9c6-cf32-45b3-8031-f8c787649c42","Type":"ContainerDied","Data":"3e60c6de61ceab13fb59c89f9079b7b1a86d2fa3766742c11a60190d50d680ae"} Oct 02 09:31:02 crc kubenswrapper[5035]: I1002 09:31:02.627734 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mx6nb" event={"ID":"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b","Type":"ContainerStarted","Data":"fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9"} Oct 02 09:31:02 crc kubenswrapper[5035]: I1002 09:31:02.630838 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqtkq" event={"ID":"1b6bb9c6-cf32-45b3-8031-f8c787649c42","Type":"ContainerStarted","Data":"a2235dee5c4f1f93182bd8903f97cf3ab4550d2cb5829d4e81229199003659a4"} Oct 02 09:31:02 crc kubenswrapper[5035]: I1002 09:31:02.660096 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mx6nb" podStartSLOduration=2.516879782 podStartE2EDuration="47.660065759s" podCreationTimestamp="2025-10-02 09:30:15 +0000 UTC" firstStartedPulling="2025-10-02 09:30:17.279107667 +0000 UTC m=+182.635451692" lastFinishedPulling="2025-10-02 09:31:02.422293644 +0000 UTC m=+227.778637669" observedRunningTime="2025-10-02 09:31:02.651216101 +0000 UTC m=+228.007560166" watchObservedRunningTime="2025-10-02 09:31:02.660065759 +0000 UTC m=+228.016409824" Oct 02 09:31:02 crc kubenswrapper[5035]: I1002 09:31:02.678031 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tqtkq" podStartSLOduration=2.478493745 podStartE2EDuration="50.678005013s" podCreationTimestamp="2025-10-02 09:30:12 +0000 UTC" firstStartedPulling="2025-10-02 09:30:13.931188791 +0000 UTC m=+179.287532806" lastFinishedPulling="2025-10-02 09:31:02.130700039 +0000 UTC m=+227.487044074" observedRunningTime="2025-10-02 09:31:02.675007962 +0000 UTC m=+228.031352037" watchObservedRunningTime="2025-10-02 09:31:02.678005013 +0000 UTC m=+228.034349078" Oct 02 09:31:02 crc kubenswrapper[5035]: I1002 09:31:02.955091 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:31:02 crc kubenswrapper[5035]: I1002 09:31:02.955179 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:31:04 crc kubenswrapper[5035]: I1002 09:31:04.104241 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-tqtkq" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="registry-server" probeResult="failure" output=< Oct 02 09:31:04 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Oct 02 09:31:04 crc kubenswrapper[5035]: > Oct 02 09:31:04 crc kubenswrapper[5035]: I1002 09:31:04.643128 5035 generic.go:334] "Generic (PLEG): container finished" podID="67c99626-0b6b-4255-80f9-836f764375c4" containerID="d1c120e1c70787582c27b4ca6b813f410565eeea9b15292bd69327699265d4df" exitCode=0 Oct 02 09:31:04 crc kubenswrapper[5035]: I1002 09:31:04.643208 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxgqx" event={"ID":"67c99626-0b6b-4255-80f9-836f764375c4","Type":"ContainerDied","Data":"d1c120e1c70787582c27b4ca6b813f410565eeea9b15292bd69327699265d4df"} Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.196111 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.196784 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.662444 5035 generic.go:334] "Generic (PLEG): container finished" podID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerID="f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15" exitCode=0 Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.662510 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbpds" event={"ID":"dbb7f782-87ae-4f6c-8642-520a483497c0","Type":"ContainerDied","Data":"f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15"} Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.665233 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8hgc" event={"ID":"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e","Type":"ContainerStarted","Data":"e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98"} Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.681940 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxgqx" event={"ID":"67c99626-0b6b-4255-80f9-836f764375c4","Type":"ContainerStarted","Data":"38cb501e7d50979a22f2f91e0d77e922801c9229e0f31de9e18a0a9d9b867ad0"} Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.687751 5035 generic.go:334] "Generic (PLEG): container finished" podID="09530660-adcc-4a75-8550-116b440990d5" containerID="1c5f1083747d8f366b30bb12f9f8bd70ea37f0415579b92cfcd80f73242c9ed2" exitCode=0 Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.687832 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47v4z" event={"ID":"09530660-adcc-4a75-8550-116b440990d5","Type":"ContainerDied","Data":"1c5f1083747d8f366b30bb12f9f8bd70ea37f0415579b92cfcd80f73242c9ed2"} Oct 02 09:31:06 crc kubenswrapper[5035]: I1002 09:31:06.758759 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sxgqx" podStartSLOduration=4.2999374 podStartE2EDuration="54.758742973s" podCreationTimestamp="2025-10-02 09:30:12 +0000 UTC" firstStartedPulling="2025-10-02 09:30:15.124288776 +0000 UTC m=+180.480632801" lastFinishedPulling="2025-10-02 09:31:05.583094349 +0000 UTC m=+230.939438374" observedRunningTime="2025-10-02 09:31:06.755747262 +0000 UTC m=+232.112091307" watchObservedRunningTime="2025-10-02 09:31:06.758742973 +0000 UTC m=+232.115086998" Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.242189 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mx6nb" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="registry-server" probeResult="failure" output=< Oct 02 09:31:07 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Oct 02 09:31:07 crc kubenswrapper[5035]: > Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.699813 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47v4z" event={"ID":"09530660-adcc-4a75-8550-116b440990d5","Type":"ContainerStarted","Data":"48341c5cf08368b7dbf101be14c161faa632eaa71cf5b8c1dc2d3f7881d87e8b"} Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.702018 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbpds" event={"ID":"dbb7f782-87ae-4f6c-8642-520a483497c0","Type":"ContainerStarted","Data":"9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0"} Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.703696 5035 generic.go:334] "Generic (PLEG): container finished" podID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerID="e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98" exitCode=0 Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.703776 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8hgc" event={"ID":"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e","Type":"ContainerDied","Data":"e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98"} Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.706876 5035 generic.go:334] "Generic (PLEG): container finished" podID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerID="af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6" exitCode=0 Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.706912 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfkq" event={"ID":"af1e80e7-afaf-4fa5-993d-54fb39e36285","Type":"ContainerDied","Data":"af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6"} Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.718923 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-47v4z" podStartSLOduration=2.755191439 podStartE2EDuration="53.718899677s" podCreationTimestamp="2025-10-02 09:30:14 +0000 UTC" firstStartedPulling="2025-10-02 09:30:16.257133167 +0000 UTC m=+181.613477192" lastFinishedPulling="2025-10-02 09:31:07.220841405 +0000 UTC m=+232.577185430" observedRunningTime="2025-10-02 09:31:07.715866415 +0000 UTC m=+233.072210440" watchObservedRunningTime="2025-10-02 09:31:07.718899677 +0000 UTC m=+233.075243702" Oct 02 09:31:07 crc kubenswrapper[5035]: I1002 09:31:07.752150 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nbpds" podStartSLOduration=2.16245474 podStartE2EDuration="55.752130914s" podCreationTimestamp="2025-10-02 09:30:12 +0000 UTC" firstStartedPulling="2025-10-02 09:30:13.952593231 +0000 UTC m=+179.308937256" lastFinishedPulling="2025-10-02 09:31:07.542269405 +0000 UTC m=+232.898613430" observedRunningTime="2025-10-02 09:31:07.750260067 +0000 UTC m=+233.106604092" watchObservedRunningTime="2025-10-02 09:31:07.752130914 +0000 UTC m=+233.108474939" Oct 02 09:31:08 crc kubenswrapper[5035]: I1002 09:31:08.714785 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg5zd" event={"ID":"d727546b-168e-4cd6-9ffc-ddd2623cf743","Type":"ContainerStarted","Data":"de2e3dca041078289f81dd796ad1b6b0929d5bc75354773862674dc508b8544b"} Oct 02 09:31:09 crc kubenswrapper[5035]: I1002 09:31:09.722168 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8hgc" event={"ID":"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e","Type":"ContainerStarted","Data":"924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800"} Oct 02 09:31:09 crc kubenswrapper[5035]: I1002 09:31:09.725209 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfkq" event={"ID":"af1e80e7-afaf-4fa5-993d-54fb39e36285","Type":"ContainerStarted","Data":"9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd"} Oct 02 09:31:09 crc kubenswrapper[5035]: I1002 09:31:09.727318 5035 generic.go:334] "Generic (PLEG): container finished" podID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerID="de2e3dca041078289f81dd796ad1b6b0929d5bc75354773862674dc508b8544b" exitCode=0 Oct 02 09:31:09 crc kubenswrapper[5035]: I1002 09:31:09.727353 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg5zd" event={"ID":"d727546b-168e-4cd6-9ffc-ddd2623cf743","Type":"ContainerDied","Data":"de2e3dca041078289f81dd796ad1b6b0929d5bc75354773862674dc508b8544b"} Oct 02 09:31:09 crc kubenswrapper[5035]: I1002 09:31:09.756387 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g8hgc" podStartSLOduration=2.916772911 podStartE2EDuration="54.756367853s" podCreationTimestamp="2025-10-02 09:30:15 +0000 UTC" firstStartedPulling="2025-10-02 09:30:17.255775812 +0000 UTC m=+182.612119837" lastFinishedPulling="2025-10-02 09:31:09.095370754 +0000 UTC m=+234.451714779" observedRunningTime="2025-10-02 09:31:09.752474365 +0000 UTC m=+235.108818400" watchObservedRunningTime="2025-10-02 09:31:09.756367853 +0000 UTC m=+235.112711878" Oct 02 09:31:09 crc kubenswrapper[5035]: I1002 09:31:09.786746 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-khfkq" podStartSLOduration=2.5085978989999997 podStartE2EDuration="57.786729423s" podCreationTimestamp="2025-10-02 09:30:12 +0000 UTC" firstStartedPulling="2025-10-02 09:30:13.974007031 +0000 UTC m=+179.330351056" lastFinishedPulling="2025-10-02 09:31:09.252138555 +0000 UTC m=+234.608482580" observedRunningTime="2025-10-02 09:31:09.784026791 +0000 UTC m=+235.140370826" watchObservedRunningTime="2025-10-02 09:31:09.786729423 +0000 UTC m=+235.143073448" Oct 02 09:31:10 crc kubenswrapper[5035]: I1002 09:31:10.735268 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg5zd" event={"ID":"d727546b-168e-4cd6-9ffc-ddd2623cf743","Type":"ContainerStarted","Data":"e151a24e21cdb678a197953e4f70f9a8b083146c0efdb6a54ad154981279f87e"} Oct 02 09:31:10 crc kubenswrapper[5035]: I1002 09:31:10.757298 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jg5zd" podStartSLOduration=1.5426010639999999 podStartE2EDuration="56.757280392s" podCreationTimestamp="2025-10-02 09:30:14 +0000 UTC" firstStartedPulling="2025-10-02 09:30:15.12926743 +0000 UTC m=+180.485611455" lastFinishedPulling="2025-10-02 09:31:10.343946758 +0000 UTC m=+235.700290783" observedRunningTime="2025-10-02 09:31:10.756609662 +0000 UTC m=+236.112953697" watchObservedRunningTime="2025-10-02 09:31:10.757280392 +0000 UTC m=+236.113624417" Oct 02 09:31:12 crc kubenswrapper[5035]: I1002 09:31:12.551000 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:31:12 crc kubenswrapper[5035]: I1002 09:31:12.551300 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:31:12 crc kubenswrapper[5035]: I1002 09:31:12.625802 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:31:12 crc kubenswrapper[5035]: I1002 09:31:12.761201 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:31:12 crc kubenswrapper[5035]: I1002 09:31:12.761359 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:31:12 crc kubenswrapper[5035]: I1002 09:31:12.809242 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:31:12 crc kubenswrapper[5035]: I1002 09:31:12.997034 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:31:13 crc kubenswrapper[5035]: I1002 09:31:13.038429 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:31:13 crc kubenswrapper[5035]: I1002 09:31:13.231104 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:31:13 crc kubenswrapper[5035]: I1002 09:31:13.231370 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:31:13 crc kubenswrapper[5035]: I1002 09:31:13.276378 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:31:13 crc kubenswrapper[5035]: I1002 09:31:13.605686 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqtkq"] Oct 02 09:31:13 crc kubenswrapper[5035]: I1002 09:31:13.799730 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:31:13 crc kubenswrapper[5035]: I1002 09:31:13.802947 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:31:14 crc kubenswrapper[5035]: I1002 09:31:14.544697 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:31:14 crc kubenswrapper[5035]: I1002 09:31:14.544742 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:31:14 crc kubenswrapper[5035]: I1002 09:31:14.599340 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:31:14 crc kubenswrapper[5035]: I1002 09:31:14.765927 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tqtkq" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="registry-server" containerID="cri-o://a2235dee5c4f1f93182bd8903f97cf3ab4550d2cb5829d4e81229199003659a4" gracePeriod=2 Oct 02 09:31:14 crc kubenswrapper[5035]: I1002 09:31:14.976392 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:31:14 crc kubenswrapper[5035]: I1002 09:31:14.976461 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.015375 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.753131 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.753423 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.774609 5035 generic.go:334] "Generic (PLEG): container finished" podID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerID="a2235dee5c4f1f93182bd8903f97cf3ab4550d2cb5829d4e81229199003659a4" exitCode=0 Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.775150 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqtkq" event={"ID":"1b6bb9c6-cf32-45b3-8031-f8c787649c42","Type":"ContainerDied","Data":"a2235dee5c4f1f93182bd8903f97cf3ab4550d2cb5829d4e81229199003659a4"} Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.803563 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.820412 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:31:15 crc kubenswrapper[5035]: I1002 09:31:15.847923 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.013147 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sxgqx"] Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.137496 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.251514 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.296205 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.304694 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-utilities\") pod \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.304816 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65mg2\" (UniqueName: \"kubernetes.io/projected/1b6bb9c6-cf32-45b3-8031-f8c787649c42-kube-api-access-65mg2\") pod \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.304852 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-catalog-content\") pod \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\" (UID: \"1b6bb9c6-cf32-45b3-8031-f8c787649c42\") " Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.308318 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-utilities" (OuterVolumeSpecName: "utilities") pod "1b6bb9c6-cf32-45b3-8031-f8c787649c42" (UID: "1b6bb9c6-cf32-45b3-8031-f8c787649c42"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.315682 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b6bb9c6-cf32-45b3-8031-f8c787649c42-kube-api-access-65mg2" (OuterVolumeSpecName: "kube-api-access-65mg2") pod "1b6bb9c6-cf32-45b3-8031-f8c787649c42" (UID: "1b6bb9c6-cf32-45b3-8031-f8c787649c42"). InnerVolumeSpecName "kube-api-access-65mg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.353125 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b6bb9c6-cf32-45b3-8031-f8c787649c42" (UID: "1b6bb9c6-cf32-45b3-8031-f8c787649c42"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.406664 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.406700 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65mg2\" (UniqueName: \"kubernetes.io/projected/1b6bb9c6-cf32-45b3-8031-f8c787649c42-kube-api-access-65mg2\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.406711 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b6bb9c6-cf32-45b3-8031-f8c787649c42-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.782195 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tqtkq" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.786585 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tqtkq" event={"ID":"1b6bb9c6-cf32-45b3-8031-f8c787649c42","Type":"ContainerDied","Data":"27635ce7066902204fcbdfafc862d894a5b7efefbe0e77496186d52ce83dfd16"} Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.786656 5035 scope.go:117] "RemoveContainer" containerID="a2235dee5c4f1f93182bd8903f97cf3ab4550d2cb5829d4e81229199003659a4" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.787228 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sxgqx" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="registry-server" containerID="cri-o://38cb501e7d50979a22f2f91e0d77e922801c9229e0f31de9e18a0a9d9b867ad0" gracePeriod=2 Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.802576 5035 scope.go:117] "RemoveContainer" containerID="3e60c6de61ceab13fb59c89f9079b7b1a86d2fa3766742c11a60190d50d680ae" Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.810230 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tqtkq"] Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.814352 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tqtkq"] Oct 02 09:31:16 crc kubenswrapper[5035]: I1002 09:31:16.832228 5035 scope.go:117] "RemoveContainer" containerID="a495158c37215164ea59f085a639c7609ccd9abbf1928574837db3ee27ed12a5" Oct 02 09:31:18 crc kubenswrapper[5035]: I1002 09:31:18.169744 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" path="/var/lib/kubelet/pods/1b6bb9c6-cf32-45b3-8031-f8c787649c42/volumes" Oct 02 09:31:18 crc kubenswrapper[5035]: I1002 09:31:18.412236 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47v4z"] Oct 02 09:31:18 crc kubenswrapper[5035]: I1002 09:31:18.412449 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-47v4z" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="registry-server" containerID="cri-o://48341c5cf08368b7dbf101be14c161faa632eaa71cf5b8c1dc2d3f7881d87e8b" gracePeriod=2 Oct 02 09:31:19 crc kubenswrapper[5035]: I1002 09:31:19.408059 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mx6nb"] Oct 02 09:31:19 crc kubenswrapper[5035]: I1002 09:31:19.408624 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mx6nb" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="registry-server" containerID="cri-o://fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9" gracePeriod=2 Oct 02 09:31:19 crc kubenswrapper[5035]: I1002 09:31:19.799517 5035 generic.go:334] "Generic (PLEG): container finished" podID="67c99626-0b6b-4255-80f9-836f764375c4" containerID="38cb501e7d50979a22f2f91e0d77e922801c9229e0f31de9e18a0a9d9b867ad0" exitCode=0 Oct 02 09:31:19 crc kubenswrapper[5035]: I1002 09:31:19.799573 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxgqx" event={"ID":"67c99626-0b6b-4255-80f9-836f764375c4","Type":"ContainerDied","Data":"38cb501e7d50979a22f2f91e0d77e922801c9229e0f31de9e18a0a9d9b867ad0"} Oct 02 09:31:19 crc kubenswrapper[5035]: I1002 09:31:19.802260 5035 generic.go:334] "Generic (PLEG): container finished" podID="09530660-adcc-4a75-8550-116b440990d5" containerID="48341c5cf08368b7dbf101be14c161faa632eaa71cf5b8c1dc2d3f7881d87e8b" exitCode=0 Oct 02 09:31:19 crc kubenswrapper[5035]: I1002 09:31:19.802307 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47v4z" event={"ID":"09530660-adcc-4a75-8550-116b440990d5","Type":"ContainerDied","Data":"48341c5cf08368b7dbf101be14c161faa632eaa71cf5b8c1dc2d3f7881d87e8b"} Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.010052 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.149818 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd5nv\" (UniqueName: \"kubernetes.io/projected/67c99626-0b6b-4255-80f9-836f764375c4-kube-api-access-fd5nv\") pod \"67c99626-0b6b-4255-80f9-836f764375c4\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.149980 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-utilities\") pod \"67c99626-0b6b-4255-80f9-836f764375c4\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.150012 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-catalog-content\") pod \"67c99626-0b6b-4255-80f9-836f764375c4\" (UID: \"67c99626-0b6b-4255-80f9-836f764375c4\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.152057 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-utilities" (OuterVolumeSpecName: "utilities") pod "67c99626-0b6b-4255-80f9-836f764375c4" (UID: "67c99626-0b6b-4255-80f9-836f764375c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.156707 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c99626-0b6b-4255-80f9-836f764375c4-kube-api-access-fd5nv" (OuterVolumeSpecName: "kube-api-access-fd5nv") pod "67c99626-0b6b-4255-80f9-836f764375c4" (UID: "67c99626-0b6b-4255-80f9-836f764375c4"). InnerVolumeSpecName "kube-api-access-fd5nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.237776 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67c99626-0b6b-4255-80f9-836f764375c4" (UID: "67c99626-0b6b-4255-80f9-836f764375c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.252799 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd5nv\" (UniqueName: \"kubernetes.io/projected/67c99626-0b6b-4255-80f9-836f764375c4-kube-api-access-fd5nv\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.252852 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.252867 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c99626-0b6b-4255-80f9-836f764375c4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.266010 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.353452 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-utilities\") pod \"09530660-adcc-4a75-8550-116b440990d5\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.353512 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-catalog-content\") pod \"09530660-adcc-4a75-8550-116b440990d5\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.353582 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9222\" (UniqueName: \"kubernetes.io/projected/09530660-adcc-4a75-8550-116b440990d5-kube-api-access-f9222\") pod \"09530660-adcc-4a75-8550-116b440990d5\" (UID: \"09530660-adcc-4a75-8550-116b440990d5\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.354439 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-utilities" (OuterVolumeSpecName: "utilities") pod "09530660-adcc-4a75-8550-116b440990d5" (UID: "09530660-adcc-4a75-8550-116b440990d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.358837 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09530660-adcc-4a75-8550-116b440990d5-kube-api-access-f9222" (OuterVolumeSpecName: "kube-api-access-f9222") pod "09530660-adcc-4a75-8550-116b440990d5" (UID: "09530660-adcc-4a75-8550-116b440990d5"). InnerVolumeSpecName "kube-api-access-f9222". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.374432 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09530660-adcc-4a75-8550-116b440990d5" (UID: "09530660-adcc-4a75-8550-116b440990d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.405231 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.454555 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9222\" (UniqueName: \"kubernetes.io/projected/09530660-adcc-4a75-8550-116b440990d5-kube-api-access-f9222\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.454598 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.454613 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09530660-adcc-4a75-8550-116b440990d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.555734 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-catalog-content\") pod \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.555904 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7lhj\" (UniqueName: \"kubernetes.io/projected/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-kube-api-access-x7lhj\") pod \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.556355 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-utilities\") pod \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\" (UID: \"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b\") " Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.557097 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-utilities" (OuterVolumeSpecName: "utilities") pod "a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" (UID: "a3fd37ad-729b-475d-8ff1-74e2d4a0b19b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.565204 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-kube-api-access-x7lhj" (OuterVolumeSpecName: "kube-api-access-x7lhj") pod "a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" (UID: "a3fd37ad-729b-475d-8ff1-74e2d4a0b19b"). InnerVolumeSpecName "kube-api-access-x7lhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.641704 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" (UID: "a3fd37ad-729b-475d-8ff1-74e2d4a0b19b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.657236 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.657269 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7lhj\" (UniqueName: \"kubernetes.io/projected/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-kube-api-access-x7lhj\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.657281 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.810794 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47v4z" event={"ID":"09530660-adcc-4a75-8550-116b440990d5","Type":"ContainerDied","Data":"c3d239ff5b23503971c28fefc885dae36a385edbb792b98334854ad7e1ab570a"} Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.810851 5035 scope.go:117] "RemoveContainer" containerID="48341c5cf08368b7dbf101be14c161faa632eaa71cf5b8c1dc2d3f7881d87e8b" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.810866 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47v4z" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.813064 5035 generic.go:334] "Generic (PLEG): container finished" podID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerID="fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9" exitCode=0 Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.813221 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mx6nb" event={"ID":"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b","Type":"ContainerDied","Data":"fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9"} Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.813377 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mx6nb" event={"ID":"a3fd37ad-729b-475d-8ff1-74e2d4a0b19b","Type":"ContainerDied","Data":"e78dc2c5c410dfeff72cd478b72d8433a0bc82c29ddfb5f2e5cf22ed75fd6aeb"} Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.813268 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mx6nb" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.815035 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sxgqx" event={"ID":"67c99626-0b6b-4255-80f9-836f764375c4","Type":"ContainerDied","Data":"ae0db716a12060d4a59f25da5401fe6851228a90425227ae59612fbde85b0af5"} Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.815119 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sxgqx" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.830590 5035 scope.go:117] "RemoveContainer" containerID="1c5f1083747d8f366b30bb12f9f8bd70ea37f0415579b92cfcd80f73242c9ed2" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.842789 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47v4z"] Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.844666 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-47v4z"] Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.857978 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mx6nb"] Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.861752 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mx6nb"] Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.862150 5035 scope.go:117] "RemoveContainer" containerID="d2a80cfccc6444c4d96de46e6607c43c0901fc2311d844a1de059eaee8aff6c0" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.871451 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sxgqx"] Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.875108 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sxgqx"] Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.881803 5035 scope.go:117] "RemoveContainer" containerID="fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.899800 5035 scope.go:117] "RemoveContainer" containerID="0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.914756 5035 scope.go:117] "RemoveContainer" containerID="8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.934142 5035 scope.go:117] "RemoveContainer" containerID="fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9" Oct 02 09:31:20 crc kubenswrapper[5035]: E1002 09:31:20.934604 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9\": container with ID starting with fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9 not found: ID does not exist" containerID="fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.934646 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9"} err="failed to get container status \"fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9\": rpc error: code = NotFound desc = could not find container \"fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9\": container with ID starting with fb512ea850a0fc6d990fbd7484524981c72b24ed5ef1e8f1691196169cd9e9c9 not found: ID does not exist" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.934697 5035 scope.go:117] "RemoveContainer" containerID="0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae" Oct 02 09:31:20 crc kubenswrapper[5035]: E1002 09:31:20.934951 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae\": container with ID starting with 0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae not found: ID does not exist" containerID="0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.934973 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae"} err="failed to get container status \"0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae\": rpc error: code = NotFound desc = could not find container \"0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae\": container with ID starting with 0a908b55c8c9d1d9f7daeee907b88b3b409fe6f3678f93a9674c2ecee58036ae not found: ID does not exist" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.934986 5035 scope.go:117] "RemoveContainer" containerID="8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b" Oct 02 09:31:20 crc kubenswrapper[5035]: E1002 09:31:20.935200 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b\": container with ID starting with 8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b not found: ID does not exist" containerID="8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.935220 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b"} err="failed to get container status \"8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b\": rpc error: code = NotFound desc = could not find container \"8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b\": container with ID starting with 8aabbe7cd18e179cec92ac0ea9bccfffff2f13525f1e86dc3ffb6507302dc29b not found: ID does not exist" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.935232 5035 scope.go:117] "RemoveContainer" containerID="38cb501e7d50979a22f2f91e0d77e922801c9229e0f31de9e18a0a9d9b867ad0" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.959583 5035 scope.go:117] "RemoveContainer" containerID="d1c120e1c70787582c27b4ca6b813f410565eeea9b15292bd69327699265d4df" Oct 02 09:31:20 crc kubenswrapper[5035]: I1002 09:31:20.981378 5035 scope.go:117] "RemoveContainer" containerID="0b24df4157d616cb2ac9c2ddaa6becf68eda381e7edfacf07f0735ce34036447" Oct 02 09:31:22 crc kubenswrapper[5035]: I1002 09:31:22.267724 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09530660-adcc-4a75-8550-116b440990d5" path="/var/lib/kubelet/pods/09530660-adcc-4a75-8550-116b440990d5/volumes" Oct 02 09:31:22 crc kubenswrapper[5035]: I1002 09:31:22.269090 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67c99626-0b6b-4255-80f9-836f764375c4" path="/var/lib/kubelet/pods/67c99626-0b6b-4255-80f9-836f764375c4/volumes" Oct 02 09:31:22 crc kubenswrapper[5035]: I1002 09:31:22.270974 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" path="/var/lib/kubelet/pods/a3fd37ad-729b-475d-8ff1-74e2d4a0b19b/volumes" Oct 02 09:31:22 crc kubenswrapper[5035]: I1002 09:31:22.605287 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:31:24 crc kubenswrapper[5035]: I1002 09:31:24.580701 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.537405 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.537468 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.537512 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.539758 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.539821 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193" gracePeriod=600 Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836013 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n6r9b"] Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836548 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836563 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836578 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9af7278-65f4-471e-8b74-4305e9174038" containerName="collect-profiles" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836588 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9af7278-65f4-471e-8b74-4305e9174038" containerName="collect-profiles" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836601 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1801415-437b-4db2-9a7d-dccf260a0184" containerName="pruner" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836609 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1801415-437b-4db2-9a7d-dccf260a0184" containerName="pruner" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836619 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836626 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836637 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836646 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836656 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836663 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836675 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836682 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836693 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836700 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836714 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6" containerName="pruner" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836722 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6" containerName="pruner" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836734 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836741 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836752 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836760 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="extract-content" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836772 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836782 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836793 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836801 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836810 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836818 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: E1002 09:31:25.836828 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836835 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="extract-utilities" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836936 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1801415-437b-4db2-9a7d-dccf260a0184" containerName="pruner" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836948 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="09530660-adcc-4a75-8550-116b440990d5" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836962 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b6bb9c6-cf32-45b3-8031-f8c787649c42" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836974 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7123e7-7420-4d3c-b6ad-c52dcb00c2e6" containerName="pruner" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836986 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c99626-0b6b-4255-80f9-836f764375c4" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.836996 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fd37ad-729b-475d-8ff1-74e2d4a0b19b" containerName="registry-server" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.837008 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9af7278-65f4-471e-8b74-4305e9174038" containerName="collect-profiles" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.837435 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:25 crc kubenswrapper[5035]: I1002 09:31:25.851112 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n6r9b"] Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023019 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-bound-sa-token\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023069 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-registry-tls\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023104 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxzmd\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-kube-api-access-fxzmd\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023134 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4921c7d9-9c54-4530-b24b-1c7714a75ec1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023243 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4921c7d9-9c54-4530-b24b-1c7714a75ec1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023291 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4921c7d9-9c54-4530-b24b-1c7714a75ec1-registry-certificates\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023454 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4921c7d9-9c54-4530-b24b-1c7714a75ec1-trusted-ca\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.023579 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.046249 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.124768 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4921c7d9-9c54-4530-b24b-1c7714a75ec1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.124844 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4921c7d9-9c54-4530-b24b-1c7714a75ec1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.124877 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4921c7d9-9c54-4530-b24b-1c7714a75ec1-registry-certificates\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.124923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4921c7d9-9c54-4530-b24b-1c7714a75ec1-trusted-ca\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.124973 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-bound-sa-token\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.124995 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-registry-tls\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.125030 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxzmd\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-kube-api-access-fxzmd\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.125701 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4921c7d9-9c54-4530-b24b-1c7714a75ec1-ca-trust-extracted\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.126372 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4921c7d9-9c54-4530-b24b-1c7714a75ec1-registry-certificates\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.126448 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4921c7d9-9c54-4530-b24b-1c7714a75ec1-trusted-ca\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.130865 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4921c7d9-9c54-4530-b24b-1c7714a75ec1-installation-pull-secrets\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.137288 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-registry-tls\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.140157 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-bound-sa-token\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.141163 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxzmd\" (UniqueName: \"kubernetes.io/projected/4921c7d9-9c54-4530-b24b-1c7714a75ec1-kube-api-access-fxzmd\") pod \"image-registry-66df7c8f76-n6r9b\" (UID: \"4921c7d9-9c54-4530-b24b-1c7714a75ec1\") " pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.156910 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.355558 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7wsdb"] Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.619831 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-n6r9b"] Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.850970 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193" exitCode=0 Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.851048 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193"} Oct 02 09:31:26 crc kubenswrapper[5035]: I1002 09:31:26.852947 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" event={"ID":"4921c7d9-9c54-4530-b24b-1c7714a75ec1","Type":"ContainerStarted","Data":"a8d72956a93fd0e779a33066b443958ee0cf7333205de49002bb813190603a63"} Oct 02 09:31:29 crc kubenswrapper[5035]: I1002 09:31:29.869659 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" event={"ID":"4921c7d9-9c54-4530-b24b-1c7714a75ec1","Type":"ContainerStarted","Data":"924f74d87d6baa9ae4be16e43bea84ffd2623344ef1c1801bb623dab0c76e970"} Oct 02 09:31:33 crc kubenswrapper[5035]: I1002 09:31:33.906686 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"335f24689478573d5d535ae9048ba12467b7447e4ec56a5453f5afb036e40b27"} Oct 02 09:31:34 crc kubenswrapper[5035]: I1002 09:31:34.910782 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:34 crc kubenswrapper[5035]: I1002 09:31:34.929555 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" podStartSLOduration=9.929512845 podStartE2EDuration="9.929512845s" podCreationTimestamp="2025-10-02 09:31:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:31:34.92802652 +0000 UTC m=+260.284370565" watchObservedRunningTime="2025-10-02 09:31:34.929512845 +0000 UTC m=+260.285856860" Oct 02 09:31:46 crc kubenswrapper[5035]: I1002 09:31:46.173270 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-n6r9b" Oct 02 09:31:46 crc kubenswrapper[5035]: I1002 09:31:46.251807 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2rf7w"] Oct 02 09:31:51 crc kubenswrapper[5035]: I1002 09:31:51.394718 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" podUID="8abe751e-2b7d-4c51-ab84-904e88edbbbc" containerName="oauth-openshift" containerID="cri-o://53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b" gracePeriod=15 Oct 02 09:31:51 crc kubenswrapper[5035]: I1002 09:31:51.856374 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:31:51 crc kubenswrapper[5035]: I1002 09:31:51.890167 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-85d9bf6778-f54cg"] Oct 02 09:31:51 crc kubenswrapper[5035]: E1002 09:31:51.890636 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8abe751e-2b7d-4c51-ab84-904e88edbbbc" containerName="oauth-openshift" Oct 02 09:31:51 crc kubenswrapper[5035]: I1002 09:31:51.890648 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8abe751e-2b7d-4c51-ab84-904e88edbbbc" containerName="oauth-openshift" Oct 02 09:31:51 crc kubenswrapper[5035]: I1002 09:31:51.890749 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8abe751e-2b7d-4c51-ab84-904e88edbbbc" containerName="oauth-openshift" Oct 02 09:31:51 crc kubenswrapper[5035]: I1002 09:31:51.891118 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:51 crc kubenswrapper[5035]: I1002 09:31:51.904141 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-85d9bf6778-f54cg"] Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.017198 5035 generic.go:334] "Generic (PLEG): container finished" podID="8abe751e-2b7d-4c51-ab84-904e88edbbbc" containerID="53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b" exitCode=0 Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.017249 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" event={"ID":"8abe751e-2b7d-4c51-ab84-904e88edbbbc","Type":"ContainerDied","Data":"53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b"} Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.017282 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.017296 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7wsdb" event={"ID":"8abe751e-2b7d-4c51-ab84-904e88edbbbc","Type":"ContainerDied","Data":"0ec7f1f1bf63ca5fdd985238a38421a19d274a4452944b97462a16c3575f58b2"} Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.017318 5035 scope.go:117] "RemoveContainer" containerID="53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027104 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-service-ca\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027178 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw6w9\" (UniqueName: \"kubernetes.io/projected/8abe751e-2b7d-4c51-ab84-904e88edbbbc-kube-api-access-kw6w9\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027219 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-cliconfig\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027254 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-trusted-ca-bundle\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027276 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-dir\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027305 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-policies\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027349 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-login\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027377 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-ocp-branding-template\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027363 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027410 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-idp-0-file-data\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027443 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-router-certs\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027471 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-error\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027497 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-session\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027584 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-serving-cert\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027626 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-provider-selection\") pod \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\" (UID: \"8abe751e-2b7d-4c51-ab84-904e88edbbbc\") " Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027856 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-error\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027896 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027921 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-login\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027946 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.027968 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-session\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028007 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dcbca09a-24d5-4e27-878d-ba64377b80d9-audit-dir\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028038 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028064 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-router-certs\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028089 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs6w2\" (UniqueName: \"kubernetes.io/projected/dcbca09a-24d5-4e27-878d-ba64377b80d9-kube-api-access-xs6w2\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028115 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-service-ca\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028244 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028588 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028992 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.028983 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029167 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-audit-policies\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029347 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029401 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029457 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029476 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029492 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029508 5035 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.029521 5035 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8abe751e-2b7d-4c51-ab84-904e88edbbbc-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.034782 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8abe751e-2b7d-4c51-ab84-904e88edbbbc-kube-api-access-kw6w9" (OuterVolumeSpecName: "kube-api-access-kw6w9") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "kube-api-access-kw6w9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.035088 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.035493 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.035977 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.036741 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.037015 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.037221 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.038032 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.042774 5035 scope.go:117] "RemoveContainer" containerID="53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b" Oct 02 09:31:52 crc kubenswrapper[5035]: E1002 09:31:52.043387 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b\": container with ID starting with 53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b not found: ID does not exist" containerID="53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.043480 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b"} err="failed to get container status \"53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b\": rpc error: code = NotFound desc = could not find container \"53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b\": container with ID starting with 53e7f956c88fc6924c0df0df7a68065059b04bde41cc12ba265939caa019991b not found: ID does not exist" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.045828 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "8abe751e-2b7d-4c51-ab84-904e88edbbbc" (UID: "8abe751e-2b7d-4c51-ab84-904e88edbbbc"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131225 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131368 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131414 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-error\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131469 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131518 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-login\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131615 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131680 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-session\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131795 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dcbca09a-24d5-4e27-878d-ba64377b80d9-audit-dir\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.131949 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.132341 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-router-certs\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.132778 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/dcbca09a-24d5-4e27-878d-ba64377b80d9-audit-dir\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133075 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs6w2\" (UniqueName: \"kubernetes.io/projected/dcbca09a-24d5-4e27-878d-ba64377b80d9-kube-api-access-xs6w2\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133124 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133158 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-service-ca\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133185 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-audit-policies\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133268 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133287 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133304 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133322 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133340 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133357 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133374 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133408 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133429 5035 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8abe751e-2b7d-4c51-ab84-904e88edbbbc-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.133448 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw6w9\" (UniqueName: \"kubernetes.io/projected/8abe751e-2b7d-4c51-ab84-904e88edbbbc-kube-api-access-kw6w9\") on node \"crc\" DevicePath \"\"" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.134119 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.134394 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-audit-policies\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.134578 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-service-ca\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.134767 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.136276 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.136616 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.137067 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.137830 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-login\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.138315 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-user-template-error\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.138864 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-session\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.140740 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-router-certs\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.140934 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/dcbca09a-24d5-4e27-878d-ba64377b80d9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.154406 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs6w2\" (UniqueName: \"kubernetes.io/projected/dcbca09a-24d5-4e27-878d-ba64377b80d9-kube-api-access-xs6w2\") pod \"oauth-openshift-85d9bf6778-f54cg\" (UID: \"dcbca09a-24d5-4e27-878d-ba64377b80d9\") " pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.218869 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.346968 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7wsdb"] Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.350291 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7wsdb"] Oct 02 09:31:52 crc kubenswrapper[5035]: I1002 09:31:52.403763 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-85d9bf6778-f54cg"] Oct 02 09:31:52 crc kubenswrapper[5035]: W1002 09:31:52.408759 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddcbca09a_24d5_4e27_878d_ba64377b80d9.slice/crio-0fa680f60737c3d53626bbb13c66ebe647e9b48deec2e59cdcf74d53372a774b WatchSource:0}: Error finding container 0fa680f60737c3d53626bbb13c66ebe647e9b48deec2e59cdcf74d53372a774b: Status 404 returned error can't find the container with id 0fa680f60737c3d53626bbb13c66ebe647e9b48deec2e59cdcf74d53372a774b Oct 02 09:31:53 crc kubenswrapper[5035]: I1002 09:31:53.025103 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" event={"ID":"dcbca09a-24d5-4e27-878d-ba64377b80d9","Type":"ContainerStarted","Data":"c341e0774eb9deed785b0f3f781ecf34bfd733f3eecf21d16329649586b1782f"} Oct 02 09:31:53 crc kubenswrapper[5035]: I1002 09:31:53.025460 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" event={"ID":"dcbca09a-24d5-4e27-878d-ba64377b80d9","Type":"ContainerStarted","Data":"0fa680f60737c3d53626bbb13c66ebe647e9b48deec2e59cdcf74d53372a774b"} Oct 02 09:31:53 crc kubenswrapper[5035]: I1002 09:31:53.025482 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:53 crc kubenswrapper[5035]: I1002 09:31:53.032850 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" Oct 02 09:31:53 crc kubenswrapper[5035]: I1002 09:31:53.054054 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-85d9bf6778-f54cg" podStartSLOduration=27.05403068 podStartE2EDuration="27.05403068s" podCreationTimestamp="2025-10-02 09:31:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:31:53.051283246 +0000 UTC m=+278.407627301" watchObservedRunningTime="2025-10-02 09:31:53.05403068 +0000 UTC m=+278.410374745" Oct 02 09:31:54 crc kubenswrapper[5035]: I1002 09:31:54.170281 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8abe751e-2b7d-4c51-ab84-904e88edbbbc" path="/var/lib/kubelet/pods/8abe751e-2b7d-4c51-ab84-904e88edbbbc/volumes" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.197396 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khfkq"] Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.198274 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-khfkq" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="registry-server" containerID="cri-o://9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd" gracePeriod=30 Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.207885 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbpds"] Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.208206 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nbpds" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="registry-server" containerID="cri-o://9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0" gracePeriod=30 Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.221878 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jbph8"] Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.222207 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerName="marketplace-operator" containerID="cri-o://12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73" gracePeriod=30 Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.232391 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg5zd"] Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.232779 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jg5zd" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="registry-server" containerID="cri-o://e151a24e21cdb678a197953e4f70f9a8b083146c0efdb6a54ad154981279f87e" gracePeriod=30 Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.241117 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8hgc"] Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.241394 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g8hgc" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="registry-server" containerID="cri-o://924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800" gracePeriod=30 Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.248449 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jrspr"] Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.249420 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.262554 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jrspr"] Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.303867 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95bd2b11-6306-4c82-be67-be8d03769473-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.303940 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9nm6\" (UniqueName: \"kubernetes.io/projected/95bd2b11-6306-4c82-be67-be8d03769473-kube-api-access-n9nm6\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.303993 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95bd2b11-6306-4c82-be67-be8d03769473-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.404999 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95bd2b11-6306-4c82-be67-be8d03769473-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.405386 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95bd2b11-6306-4c82-be67-be8d03769473-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.405412 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9nm6\" (UniqueName: \"kubernetes.io/projected/95bd2b11-6306-4c82-be67-be8d03769473-kube-api-access-n9nm6\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.406663 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95bd2b11-6306-4c82-be67-be8d03769473-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.413899 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/95bd2b11-6306-4c82-be67-be8d03769473-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.423469 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9nm6\" (UniqueName: \"kubernetes.io/projected/95bd2b11-6306-4c82-be67-be8d03769473-kube-api-access-n9nm6\") pod \"marketplace-operator-79b997595-jrspr\" (UID: \"95bd2b11-6306-4c82-be67-be8d03769473\") " pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.570044 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.705741 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.714093 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.719973 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.723266 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811645 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc4zg\" (UniqueName: \"kubernetes.io/projected/af1e80e7-afaf-4fa5-993d-54fb39e36285-kube-api-access-xc4zg\") pod \"af1e80e7-afaf-4fa5-993d-54fb39e36285\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811729 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-operator-metrics\") pod \"3f3cf046-2194-4792-aed0-7b08d65f6457\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811779 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-utilities\") pod \"dbb7f782-87ae-4f6c-8642-520a483497c0\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811810 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-utilities\") pod \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811840 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-catalog-content\") pod \"dbb7f782-87ae-4f6c-8642-520a483497c0\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811875 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-utilities\") pod \"af1e80e7-afaf-4fa5-993d-54fb39e36285\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811899 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9stwd\" (UniqueName: \"kubernetes.io/projected/3f3cf046-2194-4792-aed0-7b08d65f6457-kube-api-access-9stwd\") pod \"3f3cf046-2194-4792-aed0-7b08d65f6457\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811931 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-trusted-ca\") pod \"3f3cf046-2194-4792-aed0-7b08d65f6457\" (UID: \"3f3cf046-2194-4792-aed0-7b08d65f6457\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811955 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-catalog-content\") pod \"af1e80e7-afaf-4fa5-993d-54fb39e36285\" (UID: \"af1e80e7-afaf-4fa5-993d-54fb39e36285\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.811981 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-catalog-content\") pod \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.812010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvzpb\" (UniqueName: \"kubernetes.io/projected/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-kube-api-access-cvzpb\") pod \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\" (UID: \"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.812038 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnd4q\" (UniqueName: \"kubernetes.io/projected/dbb7f782-87ae-4f6c-8642-520a483497c0-kube-api-access-rnd4q\") pod \"dbb7f782-87ae-4f6c-8642-520a483497c0\" (UID: \"dbb7f782-87ae-4f6c-8642-520a483497c0\") " Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.812482 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-utilities" (OuterVolumeSpecName: "utilities") pod "dbb7f782-87ae-4f6c-8642-520a483497c0" (UID: "dbb7f782-87ae-4f6c-8642-520a483497c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.812543 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-utilities" (OuterVolumeSpecName: "utilities") pod "6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" (UID: "6d6a45a6-8f75-47ba-86d9-cb2ff35c288e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.813106 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-utilities" (OuterVolumeSpecName: "utilities") pod "af1e80e7-afaf-4fa5-993d-54fb39e36285" (UID: "af1e80e7-afaf-4fa5-993d-54fb39e36285"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.813832 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "3f3cf046-2194-4792-aed0-7b08d65f6457" (UID: "3f3cf046-2194-4792-aed0-7b08d65f6457"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.814512 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "3f3cf046-2194-4792-aed0-7b08d65f6457" (UID: "3f3cf046-2194-4792-aed0-7b08d65f6457"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.815514 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f3cf046-2194-4792-aed0-7b08d65f6457-kube-api-access-9stwd" (OuterVolumeSpecName: "kube-api-access-9stwd") pod "3f3cf046-2194-4792-aed0-7b08d65f6457" (UID: "3f3cf046-2194-4792-aed0-7b08d65f6457"). InnerVolumeSpecName "kube-api-access-9stwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.816282 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbb7f782-87ae-4f6c-8642-520a483497c0-kube-api-access-rnd4q" (OuterVolumeSpecName: "kube-api-access-rnd4q") pod "dbb7f782-87ae-4f6c-8642-520a483497c0" (UID: "dbb7f782-87ae-4f6c-8642-520a483497c0"). InnerVolumeSpecName "kube-api-access-rnd4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.824369 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1e80e7-afaf-4fa5-993d-54fb39e36285-kube-api-access-xc4zg" (OuterVolumeSpecName: "kube-api-access-xc4zg") pod "af1e80e7-afaf-4fa5-993d-54fb39e36285" (UID: "af1e80e7-afaf-4fa5-993d-54fb39e36285"). InnerVolumeSpecName "kube-api-access-xc4zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.825641 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-kube-api-access-cvzpb" (OuterVolumeSpecName: "kube-api-access-cvzpb") pod "6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" (UID: "6d6a45a6-8f75-47ba-86d9-cb2ff35c288e"). InnerVolumeSpecName "kube-api-access-cvzpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.870407 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbb7f782-87ae-4f6c-8642-520a483497c0" (UID: "dbb7f782-87ae-4f6c-8642-520a483497c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.876299 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af1e80e7-afaf-4fa5-993d-54fb39e36285" (UID: "af1e80e7-afaf-4fa5-993d-54fb39e36285"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.900395 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" (UID: "6d6a45a6-8f75-47ba-86d9-cb2ff35c288e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913394 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc4zg\" (UniqueName: \"kubernetes.io/projected/af1e80e7-afaf-4fa5-993d-54fb39e36285-kube-api-access-xc4zg\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913437 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913450 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913461 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913469 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbb7f782-87ae-4f6c-8642-520a483497c0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913477 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913485 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9stwd\" (UniqueName: \"kubernetes.io/projected/3f3cf046-2194-4792-aed0-7b08d65f6457-kube-api-access-9stwd\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913511 5035 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3f3cf046-2194-4792-aed0-7b08d65f6457-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913519 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af1e80e7-afaf-4fa5-993d-54fb39e36285-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913552 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913567 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvzpb\" (UniqueName: \"kubernetes.io/projected/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e-kube-api-access-cvzpb\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:03 crc kubenswrapper[5035]: I1002 09:32:03.913581 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnd4q\" (UniqueName: \"kubernetes.io/projected/dbb7f782-87ae-4f6c-8642-520a483497c0-kube-api-access-rnd4q\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.029612 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jrspr"] Oct 02 09:32:04 crc kubenswrapper[5035]: W1002 09:32:04.046193 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95bd2b11_6306_4c82_be67_be8d03769473.slice/crio-4683b694fe3be309428d871c608716eceb5386267e0af71c43cdbc9c6b45023a WatchSource:0}: Error finding container 4683b694fe3be309428d871c608716eceb5386267e0af71c43cdbc9c6b45023a: Status 404 returned error can't find the container with id 4683b694fe3be309428d871c608716eceb5386267e0af71c43cdbc9c6b45023a Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.105444 5035 generic.go:334] "Generic (PLEG): container finished" podID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerID="9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd" exitCode=0 Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.105501 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfkq" event={"ID":"af1e80e7-afaf-4fa5-993d-54fb39e36285","Type":"ContainerDied","Data":"9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.105541 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-khfkq" event={"ID":"af1e80e7-afaf-4fa5-993d-54fb39e36285","Type":"ContainerDied","Data":"11cae9c1df010b2b35a856efd55eb86cf4b82ba53eb627edb7e49a1608fff1da"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.105557 5035 scope.go:117] "RemoveContainer" containerID="9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.105710 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-khfkq" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.109776 5035 generic.go:334] "Generic (PLEG): container finished" podID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerID="e151a24e21cdb678a197953e4f70f9a8b083146c0efdb6a54ad154981279f87e" exitCode=0 Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.109821 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg5zd" event={"ID":"d727546b-168e-4cd6-9ffc-ddd2623cf743","Type":"ContainerDied","Data":"e151a24e21cdb678a197953e4f70f9a8b083146c0efdb6a54ad154981279f87e"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.111686 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" event={"ID":"95bd2b11-6306-4c82-be67-be8d03769473","Type":"ContainerStarted","Data":"4683b694fe3be309428d871c608716eceb5386267e0af71c43cdbc9c6b45023a"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.114526 5035 generic.go:334] "Generic (PLEG): container finished" podID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerID="9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0" exitCode=0 Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.114618 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbpds" event={"ID":"dbb7f782-87ae-4f6c-8642-520a483497c0","Type":"ContainerDied","Data":"9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.114647 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbpds" event={"ID":"dbb7f782-87ae-4f6c-8642-520a483497c0","Type":"ContainerDied","Data":"fa1cbc8c64084d4bc5f358bf9704936ac82420ca3ac331508df7740a5184d359"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.114644 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbpds" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.118420 5035 generic.go:334] "Generic (PLEG): container finished" podID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerID="924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800" exitCode=0 Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.118452 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8hgc" event={"ID":"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e","Type":"ContainerDied","Data":"924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.118481 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g8hgc" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.118484 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g8hgc" event={"ID":"6d6a45a6-8f75-47ba-86d9-cb2ff35c288e","Type":"ContainerDied","Data":"34045aec80839f15e3a5bea86eba72f20ca19123e4fb6d18c3515260fcafd91a"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.119691 5035 generic.go:334] "Generic (PLEG): container finished" podID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerID="12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73" exitCode=0 Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.119717 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" event={"ID":"3f3cf046-2194-4792-aed0-7b08d65f6457","Type":"ContainerDied","Data":"12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.119731 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" event={"ID":"3f3cf046-2194-4792-aed0-7b08d65f6457","Type":"ContainerDied","Data":"abd026cd2b3d67d9630a89502be01a1f8c8c06272323fa31ecf488569c85e2a3"} Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.119803 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-jbph8" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.160065 5035 scope.go:117] "RemoveContainer" containerID="af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.173703 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g8hgc"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.178792 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g8hgc"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.184279 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbpds"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.190037 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nbpds"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.195615 5035 scope.go:117] "RemoveContainer" containerID="6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.199274 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-khfkq"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.213413 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-khfkq"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.223466 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jbph8"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.227728 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-jbph8"] Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.232353 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.249211 5035 scope.go:117] "RemoveContainer" containerID="9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.264676 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd\": container with ID starting with 9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd not found: ID does not exist" containerID="9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.264725 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd"} err="failed to get container status \"9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd\": rpc error: code = NotFound desc = could not find container \"9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd\": container with ID starting with 9f3a5d04d915d659524b10981a30a07c369792986f58d6ff9b3327e3357412fd not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.264752 5035 scope.go:117] "RemoveContainer" containerID="af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.265996 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6\": container with ID starting with af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6 not found: ID does not exist" containerID="af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.266021 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6"} err="failed to get container status \"af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6\": rpc error: code = NotFound desc = could not find container \"af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6\": container with ID starting with af640cf1d1c222e952e43919f808a6b389346e4da7676d13113d5c52c0f3a5e6 not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.266037 5035 scope.go:117] "RemoveContainer" containerID="6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.269861 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9\": container with ID starting with 6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9 not found: ID does not exist" containerID="6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.269910 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9"} err="failed to get container status \"6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9\": rpc error: code = NotFound desc = could not find container \"6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9\": container with ID starting with 6238940771c362b1872c64aa227634243218fc80e9e85c8ae8b1dd4876c1bdb9 not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.269948 5035 scope.go:117] "RemoveContainer" containerID="9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.294612 5035 scope.go:117] "RemoveContainer" containerID="f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.313201 5035 scope.go:117] "RemoveContainer" containerID="17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.325011 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-catalog-content\") pod \"d727546b-168e-4cd6-9ffc-ddd2623cf743\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.325062 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj6nx\" (UniqueName: \"kubernetes.io/projected/d727546b-168e-4cd6-9ffc-ddd2623cf743-kube-api-access-dj6nx\") pod \"d727546b-168e-4cd6-9ffc-ddd2623cf743\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.325097 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-utilities\") pod \"d727546b-168e-4cd6-9ffc-ddd2623cf743\" (UID: \"d727546b-168e-4cd6-9ffc-ddd2623cf743\") " Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.326084 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-utilities" (OuterVolumeSpecName: "utilities") pod "d727546b-168e-4cd6-9ffc-ddd2623cf743" (UID: "d727546b-168e-4cd6-9ffc-ddd2623cf743"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.328685 5035 scope.go:117] "RemoveContainer" containerID="9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.329178 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d727546b-168e-4cd6-9ffc-ddd2623cf743-kube-api-access-dj6nx" (OuterVolumeSpecName: "kube-api-access-dj6nx") pod "d727546b-168e-4cd6-9ffc-ddd2623cf743" (UID: "d727546b-168e-4cd6-9ffc-ddd2623cf743"). InnerVolumeSpecName "kube-api-access-dj6nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.329184 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0\": container with ID starting with 9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0 not found: ID does not exist" containerID="9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.329247 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0"} err="failed to get container status \"9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0\": rpc error: code = NotFound desc = could not find container \"9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0\": container with ID starting with 9a6699a69908f8e3b93798562bbda8138e0e44d40f9fcbaa1a4f2113d3d870b0 not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.329281 5035 scope.go:117] "RemoveContainer" containerID="f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.329623 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15\": container with ID starting with f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15 not found: ID does not exist" containerID="f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.329649 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15"} err="failed to get container status \"f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15\": rpc error: code = NotFound desc = could not find container \"f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15\": container with ID starting with f27adf9f2c499a7a089b971f98bf915e7700f0103a411011cf765732c47c6a15 not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.329663 5035 scope.go:117] "RemoveContainer" containerID="17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.329832 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db\": container with ID starting with 17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db not found: ID does not exist" containerID="17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.329853 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db"} err="failed to get container status \"17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db\": rpc error: code = NotFound desc = could not find container \"17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db\": container with ID starting with 17e8ab85899a1e529bd3f0acaeb51448ff6db5a2b2a70897c7859dd20f1df8db not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.329865 5035 scope.go:117] "RemoveContainer" containerID="924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.341417 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d727546b-168e-4cd6-9ffc-ddd2623cf743" (UID: "d727546b-168e-4cd6-9ffc-ddd2623cf743"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.341434 5035 scope.go:117] "RemoveContainer" containerID="e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.352939 5035 scope.go:117] "RemoveContainer" containerID="dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.395129 5035 scope.go:117] "RemoveContainer" containerID="924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.395639 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800\": container with ID starting with 924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800 not found: ID does not exist" containerID="924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.395674 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800"} err="failed to get container status \"924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800\": rpc error: code = NotFound desc = could not find container \"924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800\": container with ID starting with 924e53b3e2ebebc92ec9c2cb6b1ecef2dc4decc937defcb5a680c5b9a6174800 not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.395696 5035 scope.go:117] "RemoveContainer" containerID="e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.395904 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98\": container with ID starting with e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98 not found: ID does not exist" containerID="e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.395929 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98"} err="failed to get container status \"e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98\": rpc error: code = NotFound desc = could not find container \"e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98\": container with ID starting with e7f03e3fc5b27c10c10114b38ddc15163c426d3f4fbfc5bb5f97effbe52e7a98 not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.395944 5035 scope.go:117] "RemoveContainer" containerID="dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.396252 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e\": container with ID starting with dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e not found: ID does not exist" containerID="dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.396278 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e"} err="failed to get container status \"dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e\": rpc error: code = NotFound desc = could not find container \"dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e\": container with ID starting with dd33e93c919d5a71fe7188e3905e2251a433a755462390644ff1e1f5e547569e not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.396290 5035 scope.go:117] "RemoveContainer" containerID="12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.408046 5035 scope.go:117] "RemoveContainer" containerID="12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73" Oct 02 09:32:04 crc kubenswrapper[5035]: E1002 09:32:04.408553 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73\": container with ID starting with 12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73 not found: ID does not exist" containerID="12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.408620 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73"} err="failed to get container status \"12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73\": rpc error: code = NotFound desc = could not find container \"12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73\": container with ID starting with 12a1ccc64b1cb626b3caee551dd0776641a181f1dc0af7b05a5fea1e812d4e73 not found: ID does not exist" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.426601 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.426634 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj6nx\" (UniqueName: \"kubernetes.io/projected/d727546b-168e-4cd6-9ffc-ddd2623cf743-kube-api-access-dj6nx\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:04 crc kubenswrapper[5035]: I1002 09:32:04.426646 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d727546b-168e-4cd6-9ffc-ddd2623cf743-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.127390 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg5zd" event={"ID":"d727546b-168e-4cd6-9ffc-ddd2623cf743","Type":"ContainerDied","Data":"56f420064518690a20000b1bdfdbceee0d3fe1a2eb281dd6bda40edf1f5cf8ef"} Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.127782 5035 scope.go:117] "RemoveContainer" containerID="e151a24e21cdb678a197953e4f70f9a8b083146c0efdb6a54ad154981279f87e" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.127418 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg5zd" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.129265 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" event={"ID":"95bd2b11-6306-4c82-be67-be8d03769473","Type":"ContainerStarted","Data":"10f95768710f7a019bd956b1350b3dd0af325fbee5b2073d8166aeb09802bcc7"} Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.129603 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.134165 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.147121 5035 scope.go:117] "RemoveContainer" containerID="de2e3dca041078289f81dd796ad1b6b0929d5bc75354773862674dc508b8544b" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.163437 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-jrspr" podStartSLOduration=2.163410864 podStartE2EDuration="2.163410864s" podCreationTimestamp="2025-10-02 09:32:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:32:05.148840489 +0000 UTC m=+290.505184514" watchObservedRunningTime="2025-10-02 09:32:05.163410864 +0000 UTC m=+290.519754889" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.167868 5035 scope.go:117] "RemoveContainer" containerID="153e61c0cb7b0fc50dc5000e828df335bd6ae03d9d6023e8d47734534cc37edd" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.169687 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg5zd"] Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.172132 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg5zd"] Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.621996 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p2xvl"] Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622474 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622527 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622596 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622609 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622662 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622678 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622698 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622711 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622767 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622783 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622804 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622866 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622898 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622957 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.622976 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.622988 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.623004 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623051 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.623068 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerName="marketplace-operator" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623082 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerName="marketplace-operator" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.623103 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623115 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="extract-content" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.623133 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623145 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="extract-utilities" Oct 02 09:32:05 crc kubenswrapper[5035]: E1002 09:32:05.623164 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623176 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623371 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623393 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623414 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623436 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" containerName="marketplace-operator" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.623454 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" containerName="registry-server" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.624772 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.632101 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.639354 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65c1e0e-ed37-4471-b369-6aa6c863288e-catalog-content\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.639414 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8pm4\" (UniqueName: \"kubernetes.io/projected/b65c1e0e-ed37-4471-b369-6aa6c863288e-kube-api-access-f8pm4\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.639442 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65c1e0e-ed37-4471-b369-6aa6c863288e-utilities\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.642113 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p2xvl"] Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.741151 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65c1e0e-ed37-4471-b369-6aa6c863288e-catalog-content\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.741258 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8pm4\" (UniqueName: \"kubernetes.io/projected/b65c1e0e-ed37-4471-b369-6aa6c863288e-kube-api-access-f8pm4\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.741312 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65c1e0e-ed37-4471-b369-6aa6c863288e-utilities\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.742004 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b65c1e0e-ed37-4471-b369-6aa6c863288e-utilities\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.742385 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b65c1e0e-ed37-4471-b369-6aa6c863288e-catalog-content\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.777056 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8pm4\" (UniqueName: \"kubernetes.io/projected/b65c1e0e-ed37-4471-b369-6aa6c863288e-kube-api-access-f8pm4\") pod \"redhat-operators-p2xvl\" (UID: \"b65c1e0e-ed37-4471-b369-6aa6c863288e\") " pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:05 crc kubenswrapper[5035]: I1002 09:32:05.957168 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:06 crc kubenswrapper[5035]: I1002 09:32:06.174758 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f3cf046-2194-4792-aed0-7b08d65f6457" path="/var/lib/kubelet/pods/3f3cf046-2194-4792-aed0-7b08d65f6457/volumes" Oct 02 09:32:06 crc kubenswrapper[5035]: I1002 09:32:06.176235 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d6a45a6-8f75-47ba-86d9-cb2ff35c288e" path="/var/lib/kubelet/pods/6d6a45a6-8f75-47ba-86d9-cb2ff35c288e/volumes" Oct 02 09:32:06 crc kubenswrapper[5035]: I1002 09:32:06.177260 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af1e80e7-afaf-4fa5-993d-54fb39e36285" path="/var/lib/kubelet/pods/af1e80e7-afaf-4fa5-993d-54fb39e36285/volumes" Oct 02 09:32:06 crc kubenswrapper[5035]: I1002 09:32:06.178884 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d727546b-168e-4cd6-9ffc-ddd2623cf743" path="/var/lib/kubelet/pods/d727546b-168e-4cd6-9ffc-ddd2623cf743/volumes" Oct 02 09:32:06 crc kubenswrapper[5035]: I1002 09:32:06.179811 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbb7f782-87ae-4f6c-8642-520a483497c0" path="/var/lib/kubelet/pods/dbb7f782-87ae-4f6c-8642-520a483497c0/volumes" Oct 02 09:32:06 crc kubenswrapper[5035]: I1002 09:32:06.414988 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p2xvl"] Oct 02 09:32:06 crc kubenswrapper[5035]: W1002 09:32:06.419952 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb65c1e0e_ed37_4471_b369_6aa6c863288e.slice/crio-c69390d7b889459194d6f50881e34424b0a00b1574837875ac6f90a57b2885ec WatchSource:0}: Error finding container c69390d7b889459194d6f50881e34424b0a00b1574837875ac6f90a57b2885ec: Status 404 returned error can't find the container with id c69390d7b889459194d6f50881e34424b0a00b1574837875ac6f90a57b2885ec Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.157292 5035 generic.go:334] "Generic (PLEG): container finished" podID="b65c1e0e-ed37-4471-b369-6aa6c863288e" containerID="f0f678db2d6cc3883ab0bc9269685adbd8ab17c1477d1f2932d8743ec3d17bf4" exitCode=0 Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.157445 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2xvl" event={"ID":"b65c1e0e-ed37-4471-b369-6aa6c863288e","Type":"ContainerDied","Data":"f0f678db2d6cc3883ab0bc9269685adbd8ab17c1477d1f2932d8743ec3d17bf4"} Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.157733 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2xvl" event={"ID":"b65c1e0e-ed37-4471-b369-6aa6c863288e","Type":"ContainerStarted","Data":"c69390d7b889459194d6f50881e34424b0a00b1574837875ac6f90a57b2885ec"} Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.430666 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9gx62"] Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.432662 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.435239 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9gx62"] Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.435406 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.465047 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31404512-afbe-4f3e-838a-2d46075439ef-catalog-content\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.465119 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31404512-afbe-4f3e-838a-2d46075439ef-utilities\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.465143 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2kcg\" (UniqueName: \"kubernetes.io/projected/31404512-afbe-4f3e-838a-2d46075439ef-kube-api-access-l2kcg\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.566525 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31404512-afbe-4f3e-838a-2d46075439ef-utilities\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.566624 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2kcg\" (UniqueName: \"kubernetes.io/projected/31404512-afbe-4f3e-838a-2d46075439ef-kube-api-access-l2kcg\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.566731 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31404512-afbe-4f3e-838a-2d46075439ef-catalog-content\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.567180 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31404512-afbe-4f3e-838a-2d46075439ef-utilities\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.567223 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31404512-afbe-4f3e-838a-2d46075439ef-catalog-content\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.587390 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2kcg\" (UniqueName: \"kubernetes.io/projected/31404512-afbe-4f3e-838a-2d46075439ef-kube-api-access-l2kcg\") pod \"certified-operators-9gx62\" (UID: \"31404512-afbe-4f3e-838a-2d46075439ef\") " pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:07 crc kubenswrapper[5035]: I1002 09:32:07.785736 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.027269 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qdmjm"] Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.032967 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.050589 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.056092 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qdmjm"] Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.073015 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-catalog-content\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.073059 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pdv2\" (UniqueName: \"kubernetes.io/projected/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-kube-api-access-7pdv2\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.073078 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-utilities\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.173887 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-catalog-content\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.173940 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pdv2\" (UniqueName: \"kubernetes.io/projected/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-kube-api-access-7pdv2\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.173957 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-utilities\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.174377 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-utilities\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.174596 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-catalog-content\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.188315 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9gx62"] Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.198515 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pdv2\" (UniqueName: \"kubernetes.io/projected/2939c1ba-1ace-4aea-b0ff-ca389a25c3e1-kube-api-access-7pdv2\") pod \"community-operators-qdmjm\" (UID: \"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1\") " pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: W1002 09:32:08.202296 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31404512_afbe_4f3e_838a_2d46075439ef.slice/crio-1f6deacbc7835a2efbc5ec4e433ae16501013fe1d4b4ec555e5e05dfb0e958a8 WatchSource:0}: Error finding container 1f6deacbc7835a2efbc5ec4e433ae16501013fe1d4b4ec555e5e05dfb0e958a8: Status 404 returned error can't find the container with id 1f6deacbc7835a2efbc5ec4e433ae16501013fe1d4b4ec555e5e05dfb0e958a8 Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.381305 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:08 crc kubenswrapper[5035]: I1002 09:32:08.774447 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qdmjm"] Oct 02 09:32:08 crc kubenswrapper[5035]: W1002 09:32:08.781386 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2939c1ba_1ace_4aea_b0ff_ca389a25c3e1.slice/crio-06d5dac4461521e5592d1428c70a004ae63a53ccccd375685c0749c9f1efcfd8 WatchSource:0}: Error finding container 06d5dac4461521e5592d1428c70a004ae63a53ccccd375685c0749c9f1efcfd8: Status 404 returned error can't find the container with id 06d5dac4461521e5592d1428c70a004ae63a53ccccd375685c0749c9f1efcfd8 Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.168217 5035 generic.go:334] "Generic (PLEG): container finished" podID="31404512-afbe-4f3e-838a-2d46075439ef" containerID="919b2e30adf2d5e0e4485891a4133a330f4e38e1249a9c814dc05c82289b2133" exitCode=0 Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.168324 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9gx62" event={"ID":"31404512-afbe-4f3e-838a-2d46075439ef","Type":"ContainerDied","Data":"919b2e30adf2d5e0e4485891a4133a330f4e38e1249a9c814dc05c82289b2133"} Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.168604 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9gx62" event={"ID":"31404512-afbe-4f3e-838a-2d46075439ef","Type":"ContainerStarted","Data":"1f6deacbc7835a2efbc5ec4e433ae16501013fe1d4b4ec555e5e05dfb0e958a8"} Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.179996 5035 generic.go:334] "Generic (PLEG): container finished" podID="b65c1e0e-ed37-4471-b369-6aa6c863288e" containerID="e829354a610c4045660c6cefbe99d88d8fbcd4bb44d0428df3f39ca2123d9690" exitCode=0 Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.180084 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2xvl" event={"ID":"b65c1e0e-ed37-4471-b369-6aa6c863288e","Type":"ContainerDied","Data":"e829354a610c4045660c6cefbe99d88d8fbcd4bb44d0428df3f39ca2123d9690"} Oct 02 09:32:09 crc kubenswrapper[5035]: E1002 09:32:09.180956 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb65c1e0e_ed37_4471_b369_6aa6c863288e.slice/crio-conmon-e829354a610c4045660c6cefbe99d88d8fbcd4bb44d0428df3f39ca2123d9690.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb65c1e0e_ed37_4471_b369_6aa6c863288e.slice/crio-e829354a610c4045660c6cefbe99d88d8fbcd4bb44d0428df3f39ca2123d9690.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.181272 5035 generic.go:334] "Generic (PLEG): container finished" podID="2939c1ba-1ace-4aea-b0ff-ca389a25c3e1" containerID="32b86eb6bdab656c11718cfb72d6dc0bdc38137a7588c4d0802033716406cff8" exitCode=0 Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.181302 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdmjm" event={"ID":"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1","Type":"ContainerDied","Data":"32b86eb6bdab656c11718cfb72d6dc0bdc38137a7588c4d0802033716406cff8"} Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.181320 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdmjm" event={"ID":"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1","Type":"ContainerStarted","Data":"06d5dac4461521e5592d1428c70a004ae63a53ccccd375685c0749c9f1efcfd8"} Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.826845 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mslb8"] Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.829000 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.833191 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.835671 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mslb8"] Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.928128 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859b8e87-5dc4-47fb-8c0b-17bce445deae-utilities\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.928200 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859b8e87-5dc4-47fb-8c0b-17bce445deae-catalog-content\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:09 crc kubenswrapper[5035]: I1002 09:32:09.928229 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whc8t\" (UniqueName: \"kubernetes.io/projected/859b8e87-5dc4-47fb-8c0b-17bce445deae-kube-api-access-whc8t\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.029300 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859b8e87-5dc4-47fb-8c0b-17bce445deae-catalog-content\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.029376 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whc8t\" (UniqueName: \"kubernetes.io/projected/859b8e87-5dc4-47fb-8c0b-17bce445deae-kube-api-access-whc8t\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.029459 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859b8e87-5dc4-47fb-8c0b-17bce445deae-utilities\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.030007 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/859b8e87-5dc4-47fb-8c0b-17bce445deae-utilities\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.030139 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/859b8e87-5dc4-47fb-8c0b-17bce445deae-catalog-content\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.053854 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whc8t\" (UniqueName: \"kubernetes.io/projected/859b8e87-5dc4-47fb-8c0b-17bce445deae-kube-api-access-whc8t\") pod \"redhat-marketplace-mslb8\" (UID: \"859b8e87-5dc4-47fb-8c0b-17bce445deae\") " pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.153017 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.195661 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9gx62" event={"ID":"31404512-afbe-4f3e-838a-2d46075439ef","Type":"ContainerStarted","Data":"110635f54bec34ab143308591a2f95b9b9f279e0fa2f8aafe1626111ab0eee3b"} Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.207196 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p2xvl" event={"ID":"b65c1e0e-ed37-4471-b369-6aa6c863288e","Type":"ContainerStarted","Data":"3f3c938f3b02edc9f6c012ddb7bb3c789356ef25b6702c12a6244d4a27c3eb91"} Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.209521 5035 generic.go:334] "Generic (PLEG): container finished" podID="2939c1ba-1ace-4aea-b0ff-ca389a25c3e1" containerID="d2eee49c1051871e48939939e3170136c3d233e37319667d9f26e65a1600e91e" exitCode=0 Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.209579 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdmjm" event={"ID":"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1","Type":"ContainerDied","Data":"d2eee49c1051871e48939939e3170136c3d233e37319667d9f26e65a1600e91e"} Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.258988 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p2xvl" podStartSLOduration=2.815584393 podStartE2EDuration="5.258966731s" podCreationTimestamp="2025-10-02 09:32:05 +0000 UTC" firstStartedPulling="2025-10-02 09:32:07.160551468 +0000 UTC m=+292.516895513" lastFinishedPulling="2025-10-02 09:32:09.603933826 +0000 UTC m=+294.960277851" observedRunningTime="2025-10-02 09:32:10.257620149 +0000 UTC m=+295.613964174" watchObservedRunningTime="2025-10-02 09:32:10.258966731 +0000 UTC m=+295.615310756" Oct 02 09:32:10 crc kubenswrapper[5035]: I1002 09:32:10.366658 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mslb8"] Oct 02 09:32:10 crc kubenswrapper[5035]: W1002 09:32:10.373728 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod859b8e87_5dc4_47fb_8c0b_17bce445deae.slice/crio-a1b5857ffc43c5079c26e6c9931b8837fe7bff3ccdf9053ff8afc9da9e5758f1 WatchSource:0}: Error finding container a1b5857ffc43c5079c26e6c9931b8837fe7bff3ccdf9053ff8afc9da9e5758f1: Status 404 returned error can't find the container with id a1b5857ffc43c5079c26e6c9931b8837fe7bff3ccdf9053ff8afc9da9e5758f1 Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.217639 5035 generic.go:334] "Generic (PLEG): container finished" podID="31404512-afbe-4f3e-838a-2d46075439ef" containerID="110635f54bec34ab143308591a2f95b9b9f279e0fa2f8aafe1626111ab0eee3b" exitCode=0 Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.217752 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9gx62" event={"ID":"31404512-afbe-4f3e-838a-2d46075439ef","Type":"ContainerDied","Data":"110635f54bec34ab143308591a2f95b9b9f279e0fa2f8aafe1626111ab0eee3b"} Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.219741 5035 generic.go:334] "Generic (PLEG): container finished" podID="859b8e87-5dc4-47fb-8c0b-17bce445deae" containerID="d96e25682cface285e6d2ff24f40f9b7d6fad3a8b29474296d3b396a26ad3a51" exitCode=0 Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.219823 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mslb8" event={"ID":"859b8e87-5dc4-47fb-8c0b-17bce445deae","Type":"ContainerDied","Data":"d96e25682cface285e6d2ff24f40f9b7d6fad3a8b29474296d3b396a26ad3a51"} Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.219911 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mslb8" event={"ID":"859b8e87-5dc4-47fb-8c0b-17bce445deae","Type":"ContainerStarted","Data":"a1b5857ffc43c5079c26e6c9931b8837fe7bff3ccdf9053ff8afc9da9e5758f1"} Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.293917 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" podUID="cc313ae8-012e-42ee-923e-29cfbb7ce593" containerName="registry" containerID="cri-o://c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765" gracePeriod=30 Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.903401 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952489 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-certificates\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952576 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2j42\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-kube-api-access-r2j42\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952609 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-tls\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952642 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-trusted-ca\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952664 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc313ae8-012e-42ee-923e-29cfbb7ce593-installation-pull-secrets\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952824 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952857 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-bound-sa-token\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.952910 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc313ae8-012e-42ee-923e-29cfbb7ce593-ca-trust-extracted\") pod \"cc313ae8-012e-42ee-923e-29cfbb7ce593\" (UID: \"cc313ae8-012e-42ee-923e-29cfbb7ce593\") " Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.953374 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.953745 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.961416 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.961836 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.965068 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-kube-api-access-r2j42" (OuterVolumeSpecName: "kube-api-access-r2j42") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "kube-api-access-r2j42". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.973371 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc313ae8-012e-42ee-923e-29cfbb7ce593-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.992459 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc313ae8-012e-42ee-923e-29cfbb7ce593-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:32:11 crc kubenswrapper[5035]: I1002 09:32:11.996318 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "cc313ae8-012e-42ee-923e-29cfbb7ce593" (UID: "cc313ae8-012e-42ee-923e-29cfbb7ce593"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.054492 5035 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.054548 5035 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cc313ae8-012e-42ee-923e-29cfbb7ce593-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.054559 5035 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.054570 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2j42\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-kube-api-access-r2j42\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.054580 5035 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cc313ae8-012e-42ee-923e-29cfbb7ce593-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.054588 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cc313ae8-012e-42ee-923e-29cfbb7ce593-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.054596 5035 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cc313ae8-012e-42ee-923e-29cfbb7ce593-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.226757 5035 generic.go:334] "Generic (PLEG): container finished" podID="859b8e87-5dc4-47fb-8c0b-17bce445deae" containerID="1b9af3ae974d15ec988d9c995c5a7cf39b43a62983b06a5f5caca01b24ebf971" exitCode=0 Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.226820 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mslb8" event={"ID":"859b8e87-5dc4-47fb-8c0b-17bce445deae","Type":"ContainerDied","Data":"1b9af3ae974d15ec988d9c995c5a7cf39b43a62983b06a5f5caca01b24ebf971"} Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.229031 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9gx62" event={"ID":"31404512-afbe-4f3e-838a-2d46075439ef","Type":"ContainerStarted","Data":"b40ea49b6a741a1f8958aeff8adba8120b683004b5c432d16afda1e216d862b4"} Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.230734 5035 generic.go:334] "Generic (PLEG): container finished" podID="cc313ae8-012e-42ee-923e-29cfbb7ce593" containerID="c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765" exitCode=0 Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.230778 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" event={"ID":"cc313ae8-012e-42ee-923e-29cfbb7ce593","Type":"ContainerDied","Data":"c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765"} Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.230805 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.230840 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-2rf7w" event={"ID":"cc313ae8-012e-42ee-923e-29cfbb7ce593","Type":"ContainerDied","Data":"ad71f63a35ee209b57ef42a0eab79889e198b5d3da438a59b9b108bdb2875563"} Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.230867 5035 scope.go:117] "RemoveContainer" containerID="c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.233611 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qdmjm" event={"ID":"2939c1ba-1ace-4aea-b0ff-ca389a25c3e1","Type":"ContainerStarted","Data":"8de269e59885a3e46c772030dceff340896ba8e3d70793e0a6a586acc254eda0"} Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.247906 5035 scope.go:117] "RemoveContainer" containerID="c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765" Oct 02 09:32:12 crc kubenswrapper[5035]: E1002 09:32:12.248981 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765\": container with ID starting with c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765 not found: ID does not exist" containerID="c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.249036 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765"} err="failed to get container status \"c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765\": rpc error: code = NotFound desc = could not find container \"c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765\": container with ID starting with c794f1a7737047162bc594f7c5530268658136616844dea4ee898f1439577765 not found: ID does not exist" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.271365 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9gx62" podStartSLOduration=2.772895117 podStartE2EDuration="5.271343099s" podCreationTimestamp="2025-10-02 09:32:07 +0000 UTC" firstStartedPulling="2025-10-02 09:32:09.169592004 +0000 UTC m=+294.525936029" lastFinishedPulling="2025-10-02 09:32:11.668039986 +0000 UTC m=+297.024384011" observedRunningTime="2025-10-02 09:32:12.270331028 +0000 UTC m=+297.626675053" watchObservedRunningTime="2025-10-02 09:32:12.271343099 +0000 UTC m=+297.627687124" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.286239 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qdmjm" podStartSLOduration=2.779482554 podStartE2EDuration="4.286212523s" podCreationTimestamp="2025-10-02 09:32:08 +0000 UTC" firstStartedPulling="2025-10-02 09:32:09.182502609 +0000 UTC m=+294.538846634" lastFinishedPulling="2025-10-02 09:32:10.689232578 +0000 UTC m=+296.045576603" observedRunningTime="2025-10-02 09:32:12.285812381 +0000 UTC m=+297.642156406" watchObservedRunningTime="2025-10-02 09:32:12.286212523 +0000 UTC m=+297.642556548" Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.302138 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2rf7w"] Oct 02 09:32:12 crc kubenswrapper[5035]: I1002 09:32:12.304514 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-2rf7w"] Oct 02 09:32:13 crc kubenswrapper[5035]: I1002 09:32:13.243197 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mslb8" event={"ID":"859b8e87-5dc4-47fb-8c0b-17bce445deae","Type":"ContainerStarted","Data":"01045ca4bb8f0daf12c4790b0bdb74f16e1bb681cd86e3aad07431b67d34631f"} Oct 02 09:32:13 crc kubenswrapper[5035]: I1002 09:32:13.274601 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mslb8" podStartSLOduration=2.778332215 podStartE2EDuration="4.274580613s" podCreationTimestamp="2025-10-02 09:32:09 +0000 UTC" firstStartedPulling="2025-10-02 09:32:11.220875782 +0000 UTC m=+296.577219807" lastFinishedPulling="2025-10-02 09:32:12.71712418 +0000 UTC m=+298.073468205" observedRunningTime="2025-10-02 09:32:13.272629323 +0000 UTC m=+298.628973358" watchObservedRunningTime="2025-10-02 09:32:13.274580613 +0000 UTC m=+298.630924628" Oct 02 09:32:14 crc kubenswrapper[5035]: I1002 09:32:14.171291 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc313ae8-012e-42ee-923e-29cfbb7ce593" path="/var/lib/kubelet/pods/cc313ae8-012e-42ee-923e-29cfbb7ce593/volumes" Oct 02 09:32:15 crc kubenswrapper[5035]: I1002 09:32:15.958787 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:15 crc kubenswrapper[5035]: I1002 09:32:15.960519 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:15 crc kubenswrapper[5035]: I1002 09:32:15.996619 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:16 crc kubenswrapper[5035]: I1002 09:32:16.301560 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p2xvl" Oct 02 09:32:17 crc kubenswrapper[5035]: I1002 09:32:17.786809 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:17 crc kubenswrapper[5035]: I1002 09:32:17.786891 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:17 crc kubenswrapper[5035]: I1002 09:32:17.845294 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:18 crc kubenswrapper[5035]: I1002 09:32:18.308204 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9gx62" Oct 02 09:32:18 crc kubenswrapper[5035]: I1002 09:32:18.382115 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:18 crc kubenswrapper[5035]: I1002 09:32:18.382178 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:18 crc kubenswrapper[5035]: I1002 09:32:18.429149 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:19 crc kubenswrapper[5035]: I1002 09:32:19.318505 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qdmjm" Oct 02 09:32:20 crc kubenswrapper[5035]: I1002 09:32:20.154252 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:20 crc kubenswrapper[5035]: I1002 09:32:20.154309 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:20 crc kubenswrapper[5035]: I1002 09:32:20.202020 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:32:20 crc kubenswrapper[5035]: I1002 09:32:20.328442 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mslb8" Oct 02 09:33:55 crc kubenswrapper[5035]: I1002 09:33:55.537854 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:33:55 crc kubenswrapper[5035]: I1002 09:33:55.538308 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:34:25 crc kubenswrapper[5035]: I1002 09:34:25.537952 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:34:25 crc kubenswrapper[5035]: I1002 09:34:25.538556 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:34:55 crc kubenswrapper[5035]: I1002 09:34:55.537589 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:34:55 crc kubenswrapper[5035]: I1002 09:34:55.538128 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:34:55 crc kubenswrapper[5035]: I1002 09:34:55.538183 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:34:55 crc kubenswrapper[5035]: I1002 09:34:55.538725 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"335f24689478573d5d535ae9048ba12467b7447e4ec56a5453f5afb036e40b27"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:34:55 crc kubenswrapper[5035]: I1002 09:34:55.538793 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://335f24689478573d5d535ae9048ba12467b7447e4ec56a5453f5afb036e40b27" gracePeriod=600 Oct 02 09:34:56 crc kubenswrapper[5035]: I1002 09:34:56.185377 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="335f24689478573d5d535ae9048ba12467b7447e4ec56a5453f5afb036e40b27" exitCode=0 Oct 02 09:34:56 crc kubenswrapper[5035]: I1002 09:34:56.185473 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"335f24689478573d5d535ae9048ba12467b7447e4ec56a5453f5afb036e40b27"} Oct 02 09:34:56 crc kubenswrapper[5035]: I1002 09:34:56.185991 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"e2879dc2fe72fc93981ca75e0f488799e1bf9edefc894dd701b9e1403934e161"} Oct 02 09:34:56 crc kubenswrapper[5035]: I1002 09:34:56.186013 5035 scope.go:117] "RemoveContainer" containerID="d656ba8505a513ffd640a367e3d76c7375dc866e359b7243432b7787c7397193" Oct 02 09:36:55 crc kubenswrapper[5035]: I1002 09:36:55.537688 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:36:55 crc kubenswrapper[5035]: I1002 09:36:55.538706 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:37:25 crc kubenswrapper[5035]: I1002 09:37:25.537788 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:37:25 crc kubenswrapper[5035]: I1002 09:37:25.538517 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:37:55 crc kubenswrapper[5035]: I1002 09:37:55.538466 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:37:55 crc kubenswrapper[5035]: I1002 09:37:55.539010 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:37:55 crc kubenswrapper[5035]: I1002 09:37:55.539077 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:37:55 crc kubenswrapper[5035]: I1002 09:37:55.539622 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e2879dc2fe72fc93981ca75e0f488799e1bf9edefc894dd701b9e1403934e161"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:37:55 crc kubenswrapper[5035]: I1002 09:37:55.539680 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://e2879dc2fe72fc93981ca75e0f488799e1bf9edefc894dd701b9e1403934e161" gracePeriod=600 Oct 02 09:37:56 crc kubenswrapper[5035]: I1002 09:37:56.248728 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="e2879dc2fe72fc93981ca75e0f488799e1bf9edefc894dd701b9e1403934e161" exitCode=0 Oct 02 09:37:56 crc kubenswrapper[5035]: I1002 09:37:56.248802 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"e2879dc2fe72fc93981ca75e0f488799e1bf9edefc894dd701b9e1403934e161"} Oct 02 09:37:56 crc kubenswrapper[5035]: I1002 09:37:56.249153 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"f4d0414702a1d48c4861b01b90ad79bd9fac74730456a0cf8ab31e26b08e9e73"} Oct 02 09:37:56 crc kubenswrapper[5035]: I1002 09:37:56.249176 5035 scope.go:117] "RemoveContainer" containerID="335f24689478573d5d535ae9048ba12467b7447e4ec56a5453f5afb036e40b27" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.193085 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-86j6v"] Oct 02 09:38:07 crc kubenswrapper[5035]: E1002 09:38:07.194064 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc313ae8-012e-42ee-923e-29cfbb7ce593" containerName="registry" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.194082 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc313ae8-012e-42ee-923e-29cfbb7ce593" containerName="registry" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.194199 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc313ae8-012e-42ee-923e-29cfbb7ce593" containerName="registry" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.194693 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.197327 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.197327 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.197482 5035 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-x4lpc" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.205124 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-86j6v"] Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.209055 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-plbbh"] Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.209709 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-plbbh" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.212626 5035 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-vvcdc" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.221144 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-p2tjr"] Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.221833 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.223518 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-plbbh"] Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.223989 5035 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-5zbcn" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.237614 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-p2tjr"] Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.376565 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qffst\" (UniqueName: \"kubernetes.io/projected/ab058dbd-cd9f-4754-a021-d16a6351fac2-kube-api-access-qffst\") pod \"cert-manager-5b446d88c5-plbbh\" (UID: \"ab058dbd-cd9f-4754-a021-d16a6351fac2\") " pod="cert-manager/cert-manager-5b446d88c5-plbbh" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.376636 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7q97\" (UniqueName: \"kubernetes.io/projected/3af23dbf-8290-4edc-b019-e17fcffb8489-kube-api-access-r7q97\") pod \"cert-manager-webhook-5655c58dd6-p2tjr\" (UID: \"3af23dbf-8290-4edc-b019-e17fcffb8489\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.376673 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk7f8\" (UniqueName: \"kubernetes.io/projected/d2cd272f-9d31-4957-80b7-977c88b98ad0-kube-api-access-tk7f8\") pod \"cert-manager-cainjector-7f985d654d-86j6v\" (UID: \"d2cd272f-9d31-4957-80b7-977c88b98ad0\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.478052 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qffst\" (UniqueName: \"kubernetes.io/projected/ab058dbd-cd9f-4754-a021-d16a6351fac2-kube-api-access-qffst\") pod \"cert-manager-5b446d88c5-plbbh\" (UID: \"ab058dbd-cd9f-4754-a021-d16a6351fac2\") " pod="cert-manager/cert-manager-5b446d88c5-plbbh" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.478113 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7q97\" (UniqueName: \"kubernetes.io/projected/3af23dbf-8290-4edc-b019-e17fcffb8489-kube-api-access-r7q97\") pod \"cert-manager-webhook-5655c58dd6-p2tjr\" (UID: \"3af23dbf-8290-4edc-b019-e17fcffb8489\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.478140 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk7f8\" (UniqueName: \"kubernetes.io/projected/d2cd272f-9d31-4957-80b7-977c88b98ad0-kube-api-access-tk7f8\") pod \"cert-manager-cainjector-7f985d654d-86j6v\" (UID: \"d2cd272f-9d31-4957-80b7-977c88b98ad0\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.504308 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qffst\" (UniqueName: \"kubernetes.io/projected/ab058dbd-cd9f-4754-a021-d16a6351fac2-kube-api-access-qffst\") pod \"cert-manager-5b446d88c5-plbbh\" (UID: \"ab058dbd-cd9f-4754-a021-d16a6351fac2\") " pod="cert-manager/cert-manager-5b446d88c5-plbbh" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.504646 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7q97\" (UniqueName: \"kubernetes.io/projected/3af23dbf-8290-4edc-b019-e17fcffb8489-kube-api-access-r7q97\") pod \"cert-manager-webhook-5655c58dd6-p2tjr\" (UID: \"3af23dbf-8290-4edc-b019-e17fcffb8489\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.511202 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk7f8\" (UniqueName: \"kubernetes.io/projected/d2cd272f-9d31-4957-80b7-977c88b98ad0-kube-api-access-tk7f8\") pod \"cert-manager-cainjector-7f985d654d-86j6v\" (UID: \"d2cd272f-9d31-4957-80b7-977c88b98ad0\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.517430 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.527978 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-plbbh" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.541569 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.745962 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-86j6v"] Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.762089 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.807264 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-p2tjr"] Oct 02 09:38:07 crc kubenswrapper[5035]: I1002 09:38:07.810002 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-plbbh"] Oct 02 09:38:07 crc kubenswrapper[5035]: W1002 09:38:07.837778 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab058dbd_cd9f_4754_a021_d16a6351fac2.slice/crio-f0b419696dc181d67dbe4229d77b21abd2140c02b5ee7fb58bf918b29e64972f WatchSource:0}: Error finding container f0b419696dc181d67dbe4229d77b21abd2140c02b5ee7fb58bf918b29e64972f: Status 404 returned error can't find the container with id f0b419696dc181d67dbe4229d77b21abd2140c02b5ee7fb58bf918b29e64972f Oct 02 09:38:08 crc kubenswrapper[5035]: I1002 09:38:08.343764 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" event={"ID":"3af23dbf-8290-4edc-b019-e17fcffb8489","Type":"ContainerStarted","Data":"d057f0840aba117591826e5c9f7452883169a8d4713f0923eb7fa90ea271e010"} Oct 02 09:38:08 crc kubenswrapper[5035]: I1002 09:38:08.347691 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-plbbh" event={"ID":"ab058dbd-cd9f-4754-a021-d16a6351fac2","Type":"ContainerStarted","Data":"f0b419696dc181d67dbe4229d77b21abd2140c02b5ee7fb58bf918b29e64972f"} Oct 02 09:38:08 crc kubenswrapper[5035]: I1002 09:38:08.348753 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" event={"ID":"d2cd272f-9d31-4957-80b7-977c88b98ad0","Type":"ContainerStarted","Data":"717e3b03ce854c873b6c59e504e149d596c6918d470c9c952ebdb6699d1dd88d"} Oct 02 09:38:13 crc kubenswrapper[5035]: I1002 09:38:13.372827 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" event={"ID":"3af23dbf-8290-4edc-b019-e17fcffb8489","Type":"ContainerStarted","Data":"741b18e201d093e955ff58163a640997d0709b86940e38281dace60709f9dc57"} Oct 02 09:38:13 crc kubenswrapper[5035]: I1002 09:38:13.373405 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" Oct 02 09:38:13 crc kubenswrapper[5035]: I1002 09:38:13.374156 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-plbbh" event={"ID":"ab058dbd-cd9f-4754-a021-d16a6351fac2","Type":"ContainerStarted","Data":"c17c777bfd97d5d8937278205519ae910f56846ee9c817714d3640e3ad6fc13f"} Oct 02 09:38:13 crc kubenswrapper[5035]: I1002 09:38:13.375195 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" event={"ID":"d2cd272f-9d31-4957-80b7-977c88b98ad0","Type":"ContainerStarted","Data":"af886aa89d1eb4a44adb886b20ef05c83aa339017154ce24fd1f365fe3d789ac"} Oct 02 09:38:13 crc kubenswrapper[5035]: I1002 09:38:13.388789 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" podStartSLOduration=1.788331924 podStartE2EDuration="6.388756317s" podCreationTimestamp="2025-10-02 09:38:07 +0000 UTC" firstStartedPulling="2025-10-02 09:38:07.832910586 +0000 UTC m=+653.189254611" lastFinishedPulling="2025-10-02 09:38:12.433334979 +0000 UTC m=+657.789679004" observedRunningTime="2025-10-02 09:38:13.386465181 +0000 UTC m=+658.742809216" watchObservedRunningTime="2025-10-02 09:38:13.388756317 +0000 UTC m=+658.745100342" Oct 02 09:38:13 crc kubenswrapper[5035]: I1002 09:38:13.403314 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-86j6v" podStartSLOduration=1.731714634 podStartE2EDuration="6.403297428s" podCreationTimestamp="2025-10-02 09:38:07 +0000 UTC" firstStartedPulling="2025-10-02 09:38:07.761751085 +0000 UTC m=+653.118095110" lastFinishedPulling="2025-10-02 09:38:12.433333879 +0000 UTC m=+657.789677904" observedRunningTime="2025-10-02 09:38:13.401920179 +0000 UTC m=+658.758264204" watchObservedRunningTime="2025-10-02 09:38:13.403297428 +0000 UTC m=+658.759641453" Oct 02 09:38:13 crc kubenswrapper[5035]: I1002 09:38:13.414285 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-plbbh" podStartSLOduration=1.061482208 podStartE2EDuration="6.414265166s" podCreationTimestamp="2025-10-02 09:38:07 +0000 UTC" firstStartedPulling="2025-10-02 09:38:07.840703382 +0000 UTC m=+653.197047417" lastFinishedPulling="2025-10-02 09:38:13.19348635 +0000 UTC m=+658.549830375" observedRunningTime="2025-10-02 09:38:13.412891126 +0000 UTC m=+658.769235161" watchObservedRunningTime="2025-10-02 09:38:13.414265166 +0000 UTC m=+658.770609191" Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.199190 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cmx5n"] Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.200111 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-controller" containerID="cri-o://a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.200420 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="sbdb" containerID="cri-o://d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.200453 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="nbdb" containerID="cri-o://34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.200483 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="northd" containerID="cri-o://19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.200556 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.200595 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-node" containerID="cri-o://c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.200625 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-acl-logging" containerID="cri-o://521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.224057 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" containerID="cri-o://46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" gracePeriod=30 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.400148 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/3.log" Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.403092 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovn-acl-logging/0.log" Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.404173 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" exitCode=143 Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.404219 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} Oct 02 09:38:17 crc kubenswrapper[5035]: I1002 09:38:17.545232 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-p2tjr" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.171389 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/3.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.174789 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovn-acl-logging/0.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.175485 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovn-controller/0.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.175985 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234204 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gfskt"] Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234395 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234406 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234417 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="northd" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234424 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="northd" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234432 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234438 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234448 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-acl-logging" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234453 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-acl-logging" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234461 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234467 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234474 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="nbdb" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234481 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="nbdb" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234489 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-node" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234495 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-node" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234500 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="sbdb" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234506 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="sbdb" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234515 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234521 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234540 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234546 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234556 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234561 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234568 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kubecfg-setup" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234574 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kubecfg-setup" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234660 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="sbdb" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234671 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234677 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234684 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="nbdb" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234693 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="northd" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234699 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234705 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-node" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234714 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234721 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovn-acl-logging" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234728 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234735 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.234814 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234821 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.234899 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" containerName="ovnkube-controller" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.236406 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.328984 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-netd\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329040 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-ovn\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329067 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-bin\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329096 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-config\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329123 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffeef997-af36-4315-8a12-46ecf69976e6-ovn-node-metrics-cert\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329157 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-kubelet\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329173 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-log-socket\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329190 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-netns\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329214 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-script-lib\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329237 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329260 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-var-lib-openvswitch\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329278 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-slash\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329298 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-systemd-units\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329313 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-openvswitch\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329333 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-systemd\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329353 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlfx7\" (UniqueName: \"kubernetes.io/projected/ffeef997-af36-4315-8a12-46ecf69976e6-kube-api-access-tlfx7\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329375 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-node-log\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329394 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-ovn-kubernetes\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329435 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-etc-openvswitch\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329454 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-env-overrides\") pod \"ffeef997-af36-4315-8a12-46ecf69976e6\" (UID: \"ffeef997-af36-4315-8a12-46ecf69976e6\") " Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329794 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329857 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329881 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329901 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-log-socket" (OuterVolumeSpecName: "log-socket") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329890 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329921 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329963 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330002 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329957 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-node-log" (OuterVolumeSpecName: "node-log") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.329974 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330024 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330031 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330089 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-slash" (OuterVolumeSpecName: "host-slash") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330104 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330091 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330489 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.330508 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.335303 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffeef997-af36-4315-8a12-46ecf69976e6-kube-api-access-tlfx7" (OuterVolumeSpecName: "kube-api-access-tlfx7") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "kube-api-access-tlfx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.335350 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ffeef997-af36-4315-8a12-46ecf69976e6-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.342400 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ffeef997-af36-4315-8a12-46ecf69976e6" (UID: "ffeef997-af36-4315-8a12-46ecf69976e6"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.412013 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovnkube-controller/3.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.414336 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovn-acl-logging/0.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.414794 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-cmx5n_ffeef997-af36-4315-8a12-46ecf69976e6/ovn-controller/0.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415315 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" exitCode=0 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415357 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" exitCode=0 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415369 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" exitCode=0 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415379 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" exitCode=0 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415391 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" exitCode=0 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415399 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" exitCode=0 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415402 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415424 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415455 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415466 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415477 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415488 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415497 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415507 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415518 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415524 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415544 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415550 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415555 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415559 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415565 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415571 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415578 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415585 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415593 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415599 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415605 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415611 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415616 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415622 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415628 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415634 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415639 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415559 5035 scope.go:117] "RemoveContainer" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415409 5035 generic.go:334] "Generic (PLEG): container finished" podID="ffeef997-af36-4315-8a12-46ecf69976e6" containerID="a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" exitCode=143 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415741 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cmx5n" event={"ID":"ffeef997-af36-4315-8a12-46ecf69976e6","Type":"ContainerDied","Data":"a19f9ec8b217f4387af40afcb84b5379cfb2c931ffcfa405d77acc72bf74787b"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415753 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415761 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415768 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415775 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415782 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415789 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415795 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415804 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415810 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.415817 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.419066 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/2.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.420366 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/1.log" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.420402 5035 generic.go:334] "Generic (PLEG): container finished" podID="5fafe5e6-83b2-447b-9379-b26e8071166b" containerID="c84ee73fc0e719fd27b4324bcb86b267021b32fdb93655c846a1fd63b1c22189" exitCode=2 Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.420427 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerDied","Data":"c84ee73fc0e719fd27b4324bcb86b267021b32fdb93655c846a1fd63b1c22189"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.420446 5035 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b"} Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.420827 5035 scope.go:117] "RemoveContainer" containerID="c84ee73fc0e719fd27b4324bcb86b267021b32fdb93655c846a1fd63b1c22189" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.421273 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8rpfb_openshift-multus(5fafe5e6-83b2-447b-9379-b26e8071166b)\"" pod="openshift-multus/multus-8rpfb" podUID="5fafe5e6-83b2-447b-9379-b26e8071166b" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431059 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-env-overrides\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431096 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-etc-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431175 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovnkube-config\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431209 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-systemd-units\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431252 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-run-ovn-kubernetes\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431270 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-cni-netd\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431286 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-slash\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431302 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431374 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c98jk\" (UniqueName: \"kubernetes.io/projected/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-kube-api-access-c98jk\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431756 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431854 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-kubelet\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431907 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-systemd\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.431977 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-log-socket\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432035 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-node-log\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432074 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-ovn\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432118 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-run-netns\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432159 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-var-lib-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432286 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-cni-bin\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432574 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovnkube-script-lib\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432728 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovn-node-metrics-cert\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432886 5035 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432913 5035 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432965 5035 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.432987 5035 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433003 5035 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433062 5035 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433089 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlfx7\" (UniqueName: \"kubernetes.io/projected/ffeef997-af36-4315-8a12-46ecf69976e6-kube-api-access-tlfx7\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433138 5035 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433160 5035 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433173 5035 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433187 5035 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433229 5035 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433243 5035 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433255 5035 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433267 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433306 5035 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ffeef997-af36-4315-8a12-46ecf69976e6-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433322 5035 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433349 5035 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433442 5035 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ffeef997-af36-4315-8a12-46ecf69976e6-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.433459 5035 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ffeef997-af36-4315-8a12-46ecf69976e6-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.438046 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.461830 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cmx5n"] Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.467304 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cmx5n"] Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.472192 5035 scope.go:117] "RemoveContainer" containerID="d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.485045 5035 scope.go:117] "RemoveContainer" containerID="34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.496371 5035 scope.go:117] "RemoveContainer" containerID="19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.511075 5035 scope.go:117] "RemoveContainer" containerID="897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.521877 5035 scope.go:117] "RemoveContainer" containerID="c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.533355 5035 scope.go:117] "RemoveContainer" containerID="521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534103 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovn-node-metrics-cert\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534138 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovnkube-script-lib\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534168 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-env-overrides\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534195 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-etc-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534228 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovnkube-config\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534295 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-systemd-units\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534323 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-run-ovn-kubernetes\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534346 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-cni-netd\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534364 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-slash\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534384 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534409 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c98jk\" (UniqueName: \"kubernetes.io/projected/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-kube-api-access-c98jk\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534441 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534462 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-kubelet\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534483 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-systemd\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534509 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-log-socket\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534559 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-node-log\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534585 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-ovn\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534611 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-run-netns\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534630 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-var-lib-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534704 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534775 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-cni-bin\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-cni-bin\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534822 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-ovn\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534841 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-systemd-units\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534839 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-var-lib-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534863 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-run-systemd\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534887 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-run-ovn-kubernetes\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534888 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-slash\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534900 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-kubelet\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.534925 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535038 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-cni-netd\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535092 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-log-socket\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535134 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-host-run-netns\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535251 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-node-log\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535323 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-etc-openvswitch\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535406 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-env-overrides\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535479 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovnkube-script-lib\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.535518 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovnkube-config\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.542228 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-ovn-node-metrics-cert\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.551308 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c98jk\" (UniqueName: \"kubernetes.io/projected/bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f-kube-api-access-c98jk\") pod \"ovnkube-node-gfskt\" (UID: \"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f\") " pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.552039 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.570399 5035 scope.go:117] "RemoveContainer" containerID="a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" Oct 02 09:38:18 crc kubenswrapper[5035]: W1002 09:38:18.579995 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd61f149_e3a8_47b2_b0b1_ceb5f9e0cc2f.slice/crio-2b9dc9616231cb78e35eb080ac83f515f4e155c3ad258d4b4a692f95f7ecb40e WatchSource:0}: Error finding container 2b9dc9616231cb78e35eb080ac83f515f4e155c3ad258d4b4a692f95f7ecb40e: Status 404 returned error can't find the container with id 2b9dc9616231cb78e35eb080ac83f515f4e155c3ad258d4b4a692f95f7ecb40e Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.585820 5035 scope.go:117] "RemoveContainer" containerID="b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.612456 5035 scope.go:117] "RemoveContainer" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.612898 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": container with ID starting with 46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de not found: ID does not exist" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.612938 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} err="failed to get container status \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": rpc error: code = NotFound desc = could not find container \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": container with ID starting with 46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.612964 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.613209 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": container with ID starting with 1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7 not found: ID does not exist" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.613246 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} err="failed to get container status \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": rpc error: code = NotFound desc = could not find container \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": container with ID starting with 1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.613264 5035 scope.go:117] "RemoveContainer" containerID="d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.613565 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": container with ID starting with d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e not found: ID does not exist" containerID="d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.613594 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} err="failed to get container status \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": rpc error: code = NotFound desc = could not find container \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": container with ID starting with d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.613608 5035 scope.go:117] "RemoveContainer" containerID="34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.613999 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": container with ID starting with 34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9 not found: ID does not exist" containerID="34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614056 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} err="failed to get container status \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": rpc error: code = NotFound desc = could not find container \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": container with ID starting with 34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614069 5035 scope.go:117] "RemoveContainer" containerID="19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.614425 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": container with ID starting with 19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92 not found: ID does not exist" containerID="19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614447 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} err="failed to get container status \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": rpc error: code = NotFound desc = could not find container \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": container with ID starting with 19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614459 5035 scope.go:117] "RemoveContainer" containerID="897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.614713 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": container with ID starting with 897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9 not found: ID does not exist" containerID="897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614733 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} err="failed to get container status \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": rpc error: code = NotFound desc = could not find container \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": container with ID starting with 897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614749 5035 scope.go:117] "RemoveContainer" containerID="c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.614947 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": container with ID starting with c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2 not found: ID does not exist" containerID="c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614970 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} err="failed to get container status \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": rpc error: code = NotFound desc = could not find container \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": container with ID starting with c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.614984 5035 scope.go:117] "RemoveContainer" containerID="521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.615198 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": container with ID starting with 521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3 not found: ID does not exist" containerID="521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.615218 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} err="failed to get container status \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": rpc error: code = NotFound desc = could not find container \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": container with ID starting with 521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.615258 5035 scope.go:117] "RemoveContainer" containerID="a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.615505 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": container with ID starting with a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213 not found: ID does not exist" containerID="a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.615551 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} err="failed to get container status \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": rpc error: code = NotFound desc = could not find container \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": container with ID starting with a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.615570 5035 scope.go:117] "RemoveContainer" containerID="b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99" Oct 02 09:38:18 crc kubenswrapper[5035]: E1002 09:38:18.615785 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": container with ID starting with b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99 not found: ID does not exist" containerID="b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.615808 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} err="failed to get container status \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": rpc error: code = NotFound desc = could not find container \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": container with ID starting with b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.615823 5035 scope.go:117] "RemoveContainer" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616099 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} err="failed to get container status \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": rpc error: code = NotFound desc = could not find container \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": container with ID starting with 46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616121 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616308 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} err="failed to get container status \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": rpc error: code = NotFound desc = could not find container \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": container with ID starting with 1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616337 5035 scope.go:117] "RemoveContainer" containerID="d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616611 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} err="failed to get container status \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": rpc error: code = NotFound desc = could not find container \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": container with ID starting with d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616636 5035 scope.go:117] "RemoveContainer" containerID="34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616843 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} err="failed to get container status \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": rpc error: code = NotFound desc = could not find container \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": container with ID starting with 34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.616868 5035 scope.go:117] "RemoveContainer" containerID="19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617100 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} err="failed to get container status \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": rpc error: code = NotFound desc = could not find container \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": container with ID starting with 19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617125 5035 scope.go:117] "RemoveContainer" containerID="897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617346 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} err="failed to get container status \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": rpc error: code = NotFound desc = could not find container \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": container with ID starting with 897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617372 5035 scope.go:117] "RemoveContainer" containerID="c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617608 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} err="failed to get container status \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": rpc error: code = NotFound desc = could not find container \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": container with ID starting with c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617631 5035 scope.go:117] "RemoveContainer" containerID="521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617875 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} err="failed to get container status \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": rpc error: code = NotFound desc = could not find container \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": container with ID starting with 521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.617891 5035 scope.go:117] "RemoveContainer" containerID="a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618120 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} err="failed to get container status \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": rpc error: code = NotFound desc = could not find container \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": container with ID starting with a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618138 5035 scope.go:117] "RemoveContainer" containerID="b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618394 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} err="failed to get container status \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": rpc error: code = NotFound desc = could not find container \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": container with ID starting with b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618418 5035 scope.go:117] "RemoveContainer" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618677 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} err="failed to get container status \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": rpc error: code = NotFound desc = could not find container \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": container with ID starting with 46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618696 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618883 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} err="failed to get container status \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": rpc error: code = NotFound desc = could not find container \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": container with ID starting with 1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.618904 5035 scope.go:117] "RemoveContainer" containerID="d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.619136 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} err="failed to get container status \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": rpc error: code = NotFound desc = could not find container \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": container with ID starting with d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.619159 5035 scope.go:117] "RemoveContainer" containerID="34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.619473 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} err="failed to get container status \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": rpc error: code = NotFound desc = could not find container \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": container with ID starting with 34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.619496 5035 scope.go:117] "RemoveContainer" containerID="19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.619937 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} err="failed to get container status \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": rpc error: code = NotFound desc = could not find container \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": container with ID starting with 19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.619967 5035 scope.go:117] "RemoveContainer" containerID="897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620196 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} err="failed to get container status \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": rpc error: code = NotFound desc = could not find container \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": container with ID starting with 897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620214 5035 scope.go:117] "RemoveContainer" containerID="c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620406 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} err="failed to get container status \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": rpc error: code = NotFound desc = could not find container \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": container with ID starting with c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620429 5035 scope.go:117] "RemoveContainer" containerID="521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620653 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} err="failed to get container status \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": rpc error: code = NotFound desc = could not find container \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": container with ID starting with 521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620672 5035 scope.go:117] "RemoveContainer" containerID="a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620909 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} err="failed to get container status \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": rpc error: code = NotFound desc = could not find container \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": container with ID starting with a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.620936 5035 scope.go:117] "RemoveContainer" containerID="b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621146 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} err="failed to get container status \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": rpc error: code = NotFound desc = could not find container \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": container with ID starting with b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621167 5035 scope.go:117] "RemoveContainer" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621383 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} err="failed to get container status \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": rpc error: code = NotFound desc = could not find container \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": container with ID starting with 46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621407 5035 scope.go:117] "RemoveContainer" containerID="1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621721 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7"} err="failed to get container status \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": rpc error: code = NotFound desc = could not find container \"1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7\": container with ID starting with 1d4f1c229b01bee5c29d6581d3438371ae57a0bc03435d0a1343d4e07e6b10f7 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621743 5035 scope.go:117] "RemoveContainer" containerID="d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621923 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e"} err="failed to get container status \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": rpc error: code = NotFound desc = could not find container \"d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e\": container with ID starting with d12e320c5b903b9a21a4637342400fe997573ef324854cb0cb8e760b01aa623e not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.621943 5035 scope.go:117] "RemoveContainer" containerID="34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622111 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9"} err="failed to get container status \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": rpc error: code = NotFound desc = could not find container \"34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9\": container with ID starting with 34469ee7f9b7aed066e8396fa2c1923c14ace9eeba29a43547f9f10502bac0a9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622128 5035 scope.go:117] "RemoveContainer" containerID="19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622290 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92"} err="failed to get container status \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": rpc error: code = NotFound desc = could not find container \"19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92\": container with ID starting with 19bbca7225e399830691a96d0702d0a3c8529b7eb5d6ddc62f9cc6ed33f61e92 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622305 5035 scope.go:117] "RemoveContainer" containerID="897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622466 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9"} err="failed to get container status \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": rpc error: code = NotFound desc = could not find container \"897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9\": container with ID starting with 897f0fb6f8bbb87718ec76ebc4bb9d20e8733cdc70f97862baefce3adea7d8c9 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622484 5035 scope.go:117] "RemoveContainer" containerID="c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622756 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2"} err="failed to get container status \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": rpc error: code = NotFound desc = could not find container \"c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2\": container with ID starting with c79c83f314463be7fb02d36b842b72c49240077124f41e5487e45c033c97aed2 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.622777 5035 scope.go:117] "RemoveContainer" containerID="521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.623011 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3"} err="failed to get container status \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": rpc error: code = NotFound desc = could not find container \"521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3\": container with ID starting with 521f184a8ba63998493684dd78ccb07d7d08d7dc63bd95523df7e9198d0ed5b3 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.623029 5035 scope.go:117] "RemoveContainer" containerID="a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.623261 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213"} err="failed to get container status \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": rpc error: code = NotFound desc = could not find container \"a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213\": container with ID starting with a7e321b50b9550adbd5720b3c66890d90d49f157913c7166d31e0a4615941213 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.623283 5035 scope.go:117] "RemoveContainer" containerID="b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.623654 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99"} err="failed to get container status \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": rpc error: code = NotFound desc = could not find container \"b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99\": container with ID starting with b4674687d70cb00a919fdb1b6081cd8202ee7c0fe46bb0efac86b6f61176bd99 not found: ID does not exist" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.623709 5035 scope.go:117] "RemoveContainer" containerID="46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de" Oct 02 09:38:18 crc kubenswrapper[5035]: I1002 09:38:18.624179 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de"} err="failed to get container status \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": rpc error: code = NotFound desc = could not find container \"46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de\": container with ID starting with 46adf81d4ea8c7b97d38e1e4336d6008180cbe4292143ca6ba1b3e69e7e206de not found: ID does not exist" Oct 02 09:38:19 crc kubenswrapper[5035]: I1002 09:38:19.426279 5035 generic.go:334] "Generic (PLEG): container finished" podID="bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f" containerID="690b2f102a18827ca27725aa01b806ec44a983048ff44c7c31025abfa08ed0e7" exitCode=0 Oct 02 09:38:19 crc kubenswrapper[5035]: I1002 09:38:19.426326 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerDied","Data":"690b2f102a18827ca27725aa01b806ec44a983048ff44c7c31025abfa08ed0e7"} Oct 02 09:38:19 crc kubenswrapper[5035]: I1002 09:38:19.426651 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"2b9dc9616231cb78e35eb080ac83f515f4e155c3ad258d4b4a692f95f7ecb40e"} Oct 02 09:38:20 crc kubenswrapper[5035]: I1002 09:38:20.169995 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffeef997-af36-4315-8a12-46ecf69976e6" path="/var/lib/kubelet/pods/ffeef997-af36-4315-8a12-46ecf69976e6/volumes" Oct 02 09:38:20 crc kubenswrapper[5035]: I1002 09:38:20.438263 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"7f549833f5be15e7492454076507476fb5cf61fc6b4b9f0b8f816ebb23db08be"} Oct 02 09:38:20 crc kubenswrapper[5035]: I1002 09:38:20.438305 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"811f65d0b8798ef7d8ab0d5da3b42fb6d461d2f24e53bf596a6c8ef05aa8e6f3"} Oct 02 09:38:20 crc kubenswrapper[5035]: I1002 09:38:20.438314 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"70b2690d28faa8e89710a7b46d5b1808c9e1005c4a199337cc53652e4bc2adc8"} Oct 02 09:38:20 crc kubenswrapper[5035]: I1002 09:38:20.438324 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"60c39df3174da1cae381c86267d838901ef34ea2d8bd986b0429dcfa95f2d3d8"} Oct 02 09:38:20 crc kubenswrapper[5035]: I1002 09:38:20.438331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"6c690f744238984c5c01749f24670489a2a9a65ee0d42ef7c60dcf6977ae185b"} Oct 02 09:38:20 crc kubenswrapper[5035]: I1002 09:38:20.438338 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"6de5be2e79258420bada357026f2d3cf2ad54340e843cfe26e86ef5de49e7b78"} Oct 02 09:38:22 crc kubenswrapper[5035]: I1002 09:38:22.453916 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"39ae966c739021d516042515ec5061b568800a24c1746dec41a6c81f2ca657b5"} Oct 02 09:38:25 crc kubenswrapper[5035]: I1002 09:38:25.473760 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" event={"ID":"bd61f149-e3a8-47b2-b0b1-ceb5f9e0cc2f","Type":"ContainerStarted","Data":"633fe411f46071d344e402665e6137321e9b7ec46210b71e8521cbdb225566c6"} Oct 02 09:38:25 crc kubenswrapper[5035]: I1002 09:38:25.474244 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:25 crc kubenswrapper[5035]: I1002 09:38:25.474258 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:25 crc kubenswrapper[5035]: I1002 09:38:25.496237 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:25 crc kubenswrapper[5035]: I1002 09:38:25.502353 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" podStartSLOduration=7.5023359020000004 podStartE2EDuration="7.502335902s" podCreationTimestamp="2025-10-02 09:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:38:25.501147248 +0000 UTC m=+670.857491283" watchObservedRunningTime="2025-10-02 09:38:25.502335902 +0000 UTC m=+670.858679927" Oct 02 09:38:26 crc kubenswrapper[5035]: I1002 09:38:26.478754 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:26 crc kubenswrapper[5035]: I1002 09:38:26.506264 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:32 crc kubenswrapper[5035]: I1002 09:38:32.163348 5035 scope.go:117] "RemoveContainer" containerID="c84ee73fc0e719fd27b4324bcb86b267021b32fdb93655c846a1fd63b1c22189" Oct 02 09:38:32 crc kubenswrapper[5035]: E1002 09:38:32.164082 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8rpfb_openshift-multus(5fafe5e6-83b2-447b-9379-b26e8071166b)\"" pod="openshift-multus/multus-8rpfb" podUID="5fafe5e6-83b2-447b-9379-b26e8071166b" Oct 02 09:38:45 crc kubenswrapper[5035]: I1002 09:38:45.163238 5035 scope.go:117] "RemoveContainer" containerID="c84ee73fc0e719fd27b4324bcb86b267021b32fdb93655c846a1fd63b1c22189" Oct 02 09:38:45 crc kubenswrapper[5035]: I1002 09:38:45.580446 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/2.log" Oct 02 09:38:45 crc kubenswrapper[5035]: I1002 09:38:45.580845 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/1.log" Oct 02 09:38:45 crc kubenswrapper[5035]: I1002 09:38:45.580903 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8rpfb" event={"ID":"5fafe5e6-83b2-447b-9379-b26e8071166b","Type":"ContainerStarted","Data":"f384c778c0763636cdf379c129e7527e9cea0eb945c490e2e183f49b46571f59"} Oct 02 09:38:48 crc kubenswrapper[5035]: I1002 09:38:48.574521 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gfskt" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.671775 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq"] Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.673063 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.674898 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.681199 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq"] Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.839605 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.839678 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frdml\" (UniqueName: \"kubernetes.io/projected/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-kube-api-access-frdml\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.839715 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.941448 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.941551 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frdml\" (UniqueName: \"kubernetes.io/projected/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-kube-api-access-frdml\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.941588 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.942002 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.942066 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.964773 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frdml\" (UniqueName: \"kubernetes.io/projected/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-kube-api-access-frdml\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:49 crc kubenswrapper[5035]: I1002 09:38:49.990617 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:50 crc kubenswrapper[5035]: I1002 09:38:50.177841 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq"] Oct 02 09:38:50 crc kubenswrapper[5035]: W1002 09:38:50.183003 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd61d4b64_e37d_4da5_9b5b_3ec51b94e790.slice/crio-37917530e12a0fb7cf340df27b138b9935dd3500250da3f566e78edfe16f20ec WatchSource:0}: Error finding container 37917530e12a0fb7cf340df27b138b9935dd3500250da3f566e78edfe16f20ec: Status 404 returned error can't find the container with id 37917530e12a0fb7cf340df27b138b9935dd3500250da3f566e78edfe16f20ec Oct 02 09:38:50 crc kubenswrapper[5035]: I1002 09:38:50.638766 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" event={"ID":"d61d4b64-e37d-4da5-9b5b-3ec51b94e790","Type":"ContainerStarted","Data":"77025475a5182f8fe5cd522bcca244fcb95a2f4669c06593b32aeb30831739a6"} Oct 02 09:38:50 crc kubenswrapper[5035]: I1002 09:38:50.639101 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" event={"ID":"d61d4b64-e37d-4da5-9b5b-3ec51b94e790","Type":"ContainerStarted","Data":"37917530e12a0fb7cf340df27b138b9935dd3500250da3f566e78edfe16f20ec"} Oct 02 09:38:52 crc kubenswrapper[5035]: I1002 09:38:52.657002 5035 generic.go:334] "Generic (PLEG): container finished" podID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerID="77025475a5182f8fe5cd522bcca244fcb95a2f4669c06593b32aeb30831739a6" exitCode=0 Oct 02 09:38:52 crc kubenswrapper[5035]: I1002 09:38:52.658629 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" event={"ID":"d61d4b64-e37d-4da5-9b5b-3ec51b94e790","Type":"ContainerDied","Data":"77025475a5182f8fe5cd522bcca244fcb95a2f4669c06593b32aeb30831739a6"} Oct 02 09:38:54 crc kubenswrapper[5035]: I1002 09:38:54.671467 5035 generic.go:334] "Generic (PLEG): container finished" podID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerID="7b8b3d5fb88fd36153f4f03fff80584d31c5b1df1a47e23b9df3e8ecae32b913" exitCode=0 Oct 02 09:38:54 crc kubenswrapper[5035]: I1002 09:38:54.671518 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" event={"ID":"d61d4b64-e37d-4da5-9b5b-3ec51b94e790","Type":"ContainerDied","Data":"7b8b3d5fb88fd36153f4f03fff80584d31c5b1df1a47e23b9df3e8ecae32b913"} Oct 02 09:38:55 crc kubenswrapper[5035]: I1002 09:38:55.680611 5035 generic.go:334] "Generic (PLEG): container finished" podID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerID="05f2f395cd406733d611128fbfa7eedb65dd9e77b9c5ee0d7bddada9dbd4bc2b" exitCode=0 Oct 02 09:38:55 crc kubenswrapper[5035]: I1002 09:38:55.680674 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" event={"ID":"d61d4b64-e37d-4da5-9b5b-3ec51b94e790","Type":"ContainerDied","Data":"05f2f395cd406733d611128fbfa7eedb65dd9e77b9c5ee0d7bddada9dbd4bc2b"} Oct 02 09:38:56 crc kubenswrapper[5035]: I1002 09:38:56.912297 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.029013 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-bundle\") pod \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.029144 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-util\") pod \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.029197 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frdml\" (UniqueName: \"kubernetes.io/projected/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-kube-api-access-frdml\") pod \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\" (UID: \"d61d4b64-e37d-4da5-9b5b-3ec51b94e790\") " Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.029648 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-bundle" (OuterVolumeSpecName: "bundle") pod "d61d4b64-e37d-4da5-9b5b-3ec51b94e790" (UID: "d61d4b64-e37d-4da5-9b5b-3ec51b94e790"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.035266 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-kube-api-access-frdml" (OuterVolumeSpecName: "kube-api-access-frdml") pod "d61d4b64-e37d-4da5-9b5b-3ec51b94e790" (UID: "d61d4b64-e37d-4da5-9b5b-3ec51b94e790"). InnerVolumeSpecName "kube-api-access-frdml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.039364 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-util" (OuterVolumeSpecName: "util") pod "d61d4b64-e37d-4da5-9b5b-3ec51b94e790" (UID: "d61d4b64-e37d-4da5-9b5b-3ec51b94e790"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.130241 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frdml\" (UniqueName: \"kubernetes.io/projected/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-kube-api-access-frdml\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.130278 5035 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.130300 5035 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d61d4b64-e37d-4da5-9b5b-3ec51b94e790-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.695665 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" event={"ID":"d61d4b64-e37d-4da5-9b5b-3ec51b94e790","Type":"ContainerDied","Data":"37917530e12a0fb7cf340df27b138b9935dd3500250da3f566e78edfe16f20ec"} Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.695702 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37917530e12a0fb7cf340df27b138b9935dd3500250da3f566e78edfe16f20ec" Oct 02 09:38:57 crc kubenswrapper[5035]: I1002 09:38:57.695730 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.229152 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-brkcf"] Oct 02 09:39:01 crc kubenswrapper[5035]: E1002 09:39:01.229699 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerName="extract" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.229714 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerName="extract" Oct 02 09:39:01 crc kubenswrapper[5035]: E1002 09:39:01.229723 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerName="pull" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.229729 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerName="pull" Oct 02 09:39:01 crc kubenswrapper[5035]: E1002 09:39:01.229745 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerName="util" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.229752 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerName="util" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.229865 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d61d4b64-e37d-4da5-9b5b-3ec51b94e790" containerName="extract" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.230310 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.232002 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.233135 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-tnrk2" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.233355 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.241093 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-brkcf"] Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.318019 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jhl9\" (UniqueName: \"kubernetes.io/projected/de005a53-14f0-49a4-9fbf-492c4b466dcd-kube-api-access-6jhl9\") pod \"nmstate-operator-858ddd8f98-brkcf\" (UID: \"de005a53-14f0-49a4-9fbf-492c4b466dcd\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.419522 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jhl9\" (UniqueName: \"kubernetes.io/projected/de005a53-14f0-49a4-9fbf-492c4b466dcd-kube-api-access-6jhl9\") pod \"nmstate-operator-858ddd8f98-brkcf\" (UID: \"de005a53-14f0-49a4-9fbf-492c4b466dcd\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.453980 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jhl9\" (UniqueName: \"kubernetes.io/projected/de005a53-14f0-49a4-9fbf-492c4b466dcd-kube-api-access-6jhl9\") pod \"nmstate-operator-858ddd8f98-brkcf\" (UID: \"de005a53-14f0-49a4-9fbf-492c4b466dcd\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.549734 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" Oct 02 09:39:01 crc kubenswrapper[5035]: I1002 09:39:01.745960 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-brkcf"] Oct 02 09:39:02 crc kubenswrapper[5035]: I1002 09:39:02.727233 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" event={"ID":"de005a53-14f0-49a4-9fbf-492c4b466dcd","Type":"ContainerStarted","Data":"2b4caaeaef42b2e96977d4d0692e2863540641c95712be33948de1da0ef738ec"} Oct 02 09:39:03 crc kubenswrapper[5035]: I1002 09:39:03.732768 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" event={"ID":"de005a53-14f0-49a4-9fbf-492c4b466dcd","Type":"ContainerStarted","Data":"31ba7aa311a8ac908f0689346569a909532b79c9b95e544bdea338d18cb95c39"} Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.277881 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-brkcf" podStartSLOduration=7.615355056 podStartE2EDuration="9.277857208s" podCreationTimestamp="2025-10-02 09:39:01 +0000 UTC" firstStartedPulling="2025-10-02 09:39:01.754969477 +0000 UTC m=+707.111313502" lastFinishedPulling="2025-10-02 09:39:03.417471629 +0000 UTC m=+708.773815654" observedRunningTime="2025-10-02 09:39:03.753479173 +0000 UTC m=+709.109823198" watchObservedRunningTime="2025-10-02 09:39:10.277857208 +0000 UTC m=+715.634201233" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.278679 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.279686 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.281987 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-kzq7b" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.290426 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.298257 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.304019 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.317377 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.329450 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.343327 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-hwqtt\" (UID: \"b0ccfcb3-f1bb-4573-8106-c730f6b7920c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.343451 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxplt\" (UniqueName: \"kubernetes.io/projected/193c7ee9-3048-41dd-8987-d6ad9ff934fc-kube-api-access-gxplt\") pod \"nmstate-metrics-fdff9cb8d-pdzhf\" (UID: \"193c7ee9-3048-41dd-8987-d6ad9ff934fc\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.343496 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mspc5\" (UniqueName: \"kubernetes.io/projected/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-kube-api-access-mspc5\") pod \"nmstate-webhook-6cdbc54649-hwqtt\" (UID: \"b0ccfcb3-f1bb-4573-8106-c730f6b7920c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.355456 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-mk96w"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.356291 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.436836 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.437484 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.443031 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-98zlf" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.443116 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.443294 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448167 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448274 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/aa6d0e0d-c7f8-4967-9142-4ef34426f311-kube-api-access-4ljpw\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448338 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-dbus-socket\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448367 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-nmstate-lock\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448397 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxplt\" (UniqueName: \"kubernetes.io/projected/193c7ee9-3048-41dd-8987-d6ad9ff934fc-kube-api-access-gxplt\") pod \"nmstate-metrics-fdff9cb8d-pdzhf\" (UID: \"193c7ee9-3048-41dd-8987-d6ad9ff934fc\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448444 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mspc5\" (UniqueName: \"kubernetes.io/projected/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-kube-api-access-mspc5\") pod \"nmstate-webhook-6cdbc54649-hwqtt\" (UID: \"b0ccfcb3-f1bb-4573-8106-c730f6b7920c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448765 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-hwqtt\" (UID: \"b0ccfcb3-f1bb-4573-8106-c730f6b7920c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.448806 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-ovs-socket\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: E1002 09:39:10.448851 5035 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 02 09:39:10 crc kubenswrapper[5035]: E1002 09:39:10.448910 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-tls-key-pair podName:b0ccfcb3-f1bb-4573-8106-c730f6b7920c nodeName:}" failed. No retries permitted until 2025-10-02 09:39:10.948889327 +0000 UTC m=+716.305233362 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-tls-key-pair") pod "nmstate-webhook-6cdbc54649-hwqtt" (UID: "b0ccfcb3-f1bb-4573-8106-c730f6b7920c") : secret "openshift-nmstate-webhook" not found Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.471056 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mspc5\" (UniqueName: \"kubernetes.io/projected/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-kube-api-access-mspc5\") pod \"nmstate-webhook-6cdbc54649-hwqtt\" (UID: \"b0ccfcb3-f1bb-4573-8106-c730f6b7920c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.471547 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxplt\" (UniqueName: \"kubernetes.io/projected/193c7ee9-3048-41dd-8987-d6ad9ff934fc-kube-api-access-gxplt\") pod \"nmstate-metrics-fdff9cb8d-pdzhf\" (UID: \"193c7ee9-3048-41dd-8987-d6ad9ff934fc\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552174 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t5bp\" (UniqueName: \"kubernetes.io/projected/5dc578e1-5285-4af1-a1f4-23f30e092e1b-kube-api-access-2t5bp\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552252 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5dc578e1-5285-4af1-a1f4-23f30e092e1b-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552289 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-ovs-socket\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552330 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/aa6d0e0d-c7f8-4967-9142-4ef34426f311-kube-api-access-4ljpw\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552370 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5dc578e1-5285-4af1-a1f4-23f30e092e1b-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552385 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-ovs-socket\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552401 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-dbus-socket\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552432 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-nmstate-lock\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552513 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-nmstate-lock\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.552660 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/aa6d0e0d-c7f8-4967-9142-4ef34426f311-dbus-socket\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.579522 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ljpw\" (UniqueName: \"kubernetes.io/projected/aa6d0e0d-c7f8-4967-9142-4ef34426f311-kube-api-access-4ljpw\") pod \"nmstate-handler-mk96w\" (UID: \"aa6d0e0d-c7f8-4967-9142-4ef34426f311\") " pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.618600 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.624383 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-85c5b596d5-g4dmx"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.625597 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.653167 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5dc578e1-5285-4af1-a1f4-23f30e092e1b-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.654239 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t5bp\" (UniqueName: \"kubernetes.io/projected/5dc578e1-5285-4af1-a1f4-23f30e092e1b-kube-api-access-2t5bp\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.654452 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5dc578e1-5285-4af1-a1f4-23f30e092e1b-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.655414 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5dc578e1-5285-4af1-a1f4-23f30e092e1b-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.657767 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/5dc578e1-5285-4af1-a1f4-23f30e092e1b-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.675017 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.683811 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t5bp\" (UniqueName: \"kubernetes.io/projected/5dc578e1-5285-4af1-a1f4-23f30e092e1b-kube-api-access-2t5bp\") pod \"nmstate-console-plugin-6b874cbd85-5l6pn\" (UID: \"5dc578e1-5285-4af1-a1f4-23f30e092e1b\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.711093 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-85c5b596d5-g4dmx"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.756505 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-console-config\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.757129 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d76f2850-a704-48e9-b6a2-341da62dc090-console-serving-cert\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.757206 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpcb7\" (UniqueName: \"kubernetes.io/projected/d76f2850-a704-48e9-b6a2-341da62dc090-kube-api-access-jpcb7\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.757289 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-trusted-ca-bundle\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.757463 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-oauth-serving-cert\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.757513 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d76f2850-a704-48e9-b6a2-341da62dc090-console-oauth-config\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.757551 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-service-ca\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.759972 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.777113 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mk96w" event={"ID":"aa6d0e0d-c7f8-4967-9142-4ef34426f311","Type":"ContainerStarted","Data":"12593ac2a56302f07bfe99135334a17648624a23c78d850ce091d251f06d416f"} Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.859518 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-trusted-ca-bundle\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.860058 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-oauth-serving-cert\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.860095 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d76f2850-a704-48e9-b6a2-341da62dc090-console-oauth-config\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.860146 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-service-ca\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.860229 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-console-config\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.860250 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d76f2850-a704-48e9-b6a2-341da62dc090-console-serving-cert\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.860303 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpcb7\" (UniqueName: \"kubernetes.io/projected/d76f2850-a704-48e9-b6a2-341da62dc090-kube-api-access-jpcb7\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.860937 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-oauth-serving-cert\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.861622 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-console-config\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.861701 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-service-ca\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.862748 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d76f2850-a704-48e9-b6a2-341da62dc090-trusted-ca-bundle\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.864803 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d76f2850-a704-48e9-b6a2-341da62dc090-console-oauth-config\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.868010 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d76f2850-a704-48e9-b6a2-341da62dc090-console-serving-cert\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.879579 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpcb7\" (UniqueName: \"kubernetes.io/projected/d76f2850-a704-48e9-b6a2-341da62dc090-kube-api-access-jpcb7\") pod \"console-85c5b596d5-g4dmx\" (UID: \"d76f2850-a704-48e9-b6a2-341da62dc090\") " pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.879922 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.959459 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn"] Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.961086 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-hwqtt\" (UID: \"b0ccfcb3-f1bb-4573-8106-c730f6b7920c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.964094 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b0ccfcb3-f1bb-4573-8106-c730f6b7920c-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-hwqtt\" (UID: \"b0ccfcb3-f1bb-4573-8106-c730f6b7920c\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:10 crc kubenswrapper[5035]: W1002 09:39:10.965393 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5dc578e1_5285_4af1_a1f4_23f30e092e1b.slice/crio-2fd9375622eb29e8b88267e251198e51895ac4c4248268c1c625ce0086a2dde1 WatchSource:0}: Error finding container 2fd9375622eb29e8b88267e251198e51895ac4c4248268c1c625ce0086a2dde1: Status 404 returned error can't find the container with id 2fd9375622eb29e8b88267e251198e51895ac4c4248268c1c625ce0086a2dde1 Oct 02 09:39:10 crc kubenswrapper[5035]: I1002 09:39:10.983882 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.163555 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-85c5b596d5-g4dmx"] Oct 02 09:39:11 crc kubenswrapper[5035]: W1002 09:39:11.167917 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd76f2850_a704_48e9_b6a2_341da62dc090.slice/crio-901c1a128530638f2e2387ba7ab2b3f17566c953462f188b3196e471c7d58915 WatchSource:0}: Error finding container 901c1a128530638f2e2387ba7ab2b3f17566c953462f188b3196e471c7d58915: Status 404 returned error can't find the container with id 901c1a128530638f2e2387ba7ab2b3f17566c953462f188b3196e471c7d58915 Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.250763 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.444158 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt"] Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.785018 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" event={"ID":"b0ccfcb3-f1bb-4573-8106-c730f6b7920c","Type":"ContainerStarted","Data":"5b45f33e7c859a3ec7b9a18488ca0ebc87106f9ef919109abc3f678a8fe25ebc"} Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.786272 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" event={"ID":"193c7ee9-3048-41dd-8987-d6ad9ff934fc","Type":"ContainerStarted","Data":"39660a4af1c40dea0be092c3ff974593e396153a521011b5e886794a32f6b69a"} Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.787651 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" event={"ID":"5dc578e1-5285-4af1-a1f4-23f30e092e1b","Type":"ContainerStarted","Data":"2fd9375622eb29e8b88267e251198e51895ac4c4248268c1c625ce0086a2dde1"} Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.788920 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85c5b596d5-g4dmx" event={"ID":"d76f2850-a704-48e9-b6a2-341da62dc090","Type":"ContainerStarted","Data":"8940379e945e01518d781d28b2375cd37ea9b0820688013d84407940f93b1206"} Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.788983 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85c5b596d5-g4dmx" event={"ID":"d76f2850-a704-48e9-b6a2-341da62dc090","Type":"ContainerStarted","Data":"901c1a128530638f2e2387ba7ab2b3f17566c953462f188b3196e471c7d58915"} Oct 02 09:39:11 crc kubenswrapper[5035]: I1002 09:39:11.808638 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-85c5b596d5-g4dmx" podStartSLOduration=1.808622814 podStartE2EDuration="1.808622814s" podCreationTimestamp="2025-10-02 09:39:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:39:11.80609107 +0000 UTC m=+717.162435105" watchObservedRunningTime="2025-10-02 09:39:11.808622814 +0000 UTC m=+717.164966839" Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.810828 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" event={"ID":"b0ccfcb3-f1bb-4573-8106-c730f6b7920c","Type":"ContainerStarted","Data":"16caa3adc86959e3f6915dae5358b804cbb9e262bf9023225fe40ee4adc47334"} Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.811491 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.812162 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mk96w" event={"ID":"aa6d0e0d-c7f8-4967-9142-4ef34426f311","Type":"ContainerStarted","Data":"4b88125c2216db20aa51ad56afff031680b568e72dfba9591cd8a205e50548e2"} Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.812954 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.814254 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" event={"ID":"193c7ee9-3048-41dd-8987-d6ad9ff934fc","Type":"ContainerStarted","Data":"28daa9222eef1c9591ea7c1c06ceda9f6a1ef160337fbc6f79b86bffd522d677"} Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.815565 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" event={"ID":"5dc578e1-5285-4af1-a1f4-23f30e092e1b","Type":"ContainerStarted","Data":"e6cc1f66d19a9abd0b5a19c9de4c1318c983182d88194d50f73e6906c1a025fd"} Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.860351 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" podStartSLOduration=1.97374743 podStartE2EDuration="4.860334831s" podCreationTimestamp="2025-10-02 09:39:10 +0000 UTC" firstStartedPulling="2025-10-02 09:39:11.448715055 +0000 UTC m=+716.805059080" lastFinishedPulling="2025-10-02 09:39:14.335302456 +0000 UTC m=+719.691646481" observedRunningTime="2025-10-02 09:39:14.838357541 +0000 UTC m=+720.194701586" watchObservedRunningTime="2025-10-02 09:39:14.860334831 +0000 UTC m=+720.216678856" Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.860559 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-mk96w" podStartSLOduration=1.273891986 podStartE2EDuration="4.860552518s" podCreationTimestamp="2025-10-02 09:39:10 +0000 UTC" firstStartedPulling="2025-10-02 09:39:10.72306708 +0000 UTC m=+716.079411105" lastFinishedPulling="2025-10-02 09:39:14.309727612 +0000 UTC m=+719.666071637" observedRunningTime="2025-10-02 09:39:14.859826346 +0000 UTC m=+720.216170371" watchObservedRunningTime="2025-10-02 09:39:14.860552518 +0000 UTC m=+720.216896543" Oct 02 09:39:14 crc kubenswrapper[5035]: I1002 09:39:14.880933 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5l6pn" podStartSLOduration=1.551251411 podStartE2EDuration="4.88090945s" podCreationTimestamp="2025-10-02 09:39:10 +0000 UTC" firstStartedPulling="2025-10-02 09:39:10.967754024 +0000 UTC m=+716.324098049" lastFinishedPulling="2025-10-02 09:39:14.297412063 +0000 UTC m=+719.653756088" observedRunningTime="2025-10-02 09:39:14.875207334 +0000 UTC m=+720.231551359" watchObservedRunningTime="2025-10-02 09:39:14.88090945 +0000 UTC m=+720.237253475" Oct 02 09:39:16 crc kubenswrapper[5035]: I1002 09:39:16.443643 5035 scope.go:117] "RemoveContainer" containerID="bb931c90ee5bbe397858058201c7bacf79aa673a6d3c02412ed2cc82972aea9b" Oct 02 09:39:16 crc kubenswrapper[5035]: I1002 09:39:16.838550 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8rpfb_5fafe5e6-83b2-447b-9379-b26e8071166b/kube-multus/2.log" Oct 02 09:39:16 crc kubenswrapper[5035]: I1002 09:39:16.841000 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" event={"ID":"193c7ee9-3048-41dd-8987-d6ad9ff934fc","Type":"ContainerStarted","Data":"4202ac1a230e89ed411fc9d6e4a696c17e077cb207ebfe10028416032507e6fc"} Oct 02 09:39:16 crc kubenswrapper[5035]: I1002 09:39:16.863197 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-pdzhf" podStartSLOduration=1.154701236 podStartE2EDuration="6.863168472s" podCreationTimestamp="2025-10-02 09:39:10 +0000 UTC" firstStartedPulling="2025-10-02 09:39:10.88863105 +0000 UTC m=+716.244975075" lastFinishedPulling="2025-10-02 09:39:16.597098286 +0000 UTC m=+721.953442311" observedRunningTime="2025-10-02 09:39:16.857796866 +0000 UTC m=+722.214140891" watchObservedRunningTime="2025-10-02 09:39:16.863168472 +0000 UTC m=+722.219512497" Oct 02 09:39:20 crc kubenswrapper[5035]: I1002 09:39:20.699389 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-mk96w" Oct 02 09:39:20 crc kubenswrapper[5035]: I1002 09:39:20.984649 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:20 crc kubenswrapper[5035]: I1002 09:39:20.985033 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:20 crc kubenswrapper[5035]: I1002 09:39:20.990252 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:21 crc kubenswrapper[5035]: I1002 09:39:21.874734 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-85c5b596d5-g4dmx" Oct 02 09:39:21 crc kubenswrapper[5035]: I1002 09:39:21.920377 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-b7fjd"] Oct 02 09:39:31 crc kubenswrapper[5035]: I1002 09:39:31.256551 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-hwqtt" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.244088 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48"] Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.245553 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.246880 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.298989 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48"] Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.338794 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.338858 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.338877 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5jcp\" (UniqueName: \"kubernetes.io/projected/c643d696-0eaa-4495-acfe-6035dd231a0c-kube-api-access-k5jcp\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.439763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.439850 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.439873 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5jcp\" (UniqueName: \"kubernetes.io/projected/c643d696-0eaa-4495-acfe-6035dd231a0c-kube-api-access-k5jcp\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.440377 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.440383 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.468998 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5jcp\" (UniqueName: \"kubernetes.io/projected/c643d696-0eaa-4495-acfe-6035dd231a0c-kube-api-access-k5jcp\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.560723 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:45 crc kubenswrapper[5035]: I1002 09:39:45.962960 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48"] Oct 02 09:39:46 crc kubenswrapper[5035]: I1002 09:39:46.027302 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" event={"ID":"c643d696-0eaa-4495-acfe-6035dd231a0c","Type":"ContainerStarted","Data":"079135d71be8a0ad5d6b2a933ddf263bb41f9183bc4220c28dafe24b1e0d10f3"} Oct 02 09:39:46 crc kubenswrapper[5035]: I1002 09:39:46.977769 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-b7fjd" podUID="b13e03a7-b033-441d-9f0b-2c6d85f3ab25" containerName="console" containerID="cri-o://fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f" gracePeriod=15 Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.034273 5035 generic.go:334] "Generic (PLEG): container finished" podID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerID="54e607ffe47cf2de7f76f455e4906335586e187d71c520c0c9a45925c4ccca01" exitCode=0 Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.034331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" event={"ID":"c643d696-0eaa-4495-acfe-6035dd231a0c","Type":"ContainerDied","Data":"54e607ffe47cf2de7f76f455e4906335586e187d71c520c0c9a45925c4ccca01"} Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.331502 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-b7fjd_b13e03a7-b033-441d-9f0b-2c6d85f3ab25/console/0.log" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.331581 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.465095 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc9lq\" (UniqueName: \"kubernetes.io/projected/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-kube-api-access-rc9lq\") pod \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.465497 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-oauth-serving-cert\") pod \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.465590 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-trusted-ca-bundle\") pod \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.465625 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-serving-cert\") pod \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.465653 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-config\") pod \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.465702 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-oauth-config\") pod \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.465753 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-service-ca\") pod \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\" (UID: \"b13e03a7-b033-441d-9f0b-2c6d85f3ab25\") " Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.466149 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "b13e03a7-b033-441d-9f0b-2c6d85f3ab25" (UID: "b13e03a7-b033-441d-9f0b-2c6d85f3ab25"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.466180 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-config" (OuterVolumeSpecName: "console-config") pod "b13e03a7-b033-441d-9f0b-2c6d85f3ab25" (UID: "b13e03a7-b033-441d-9f0b-2c6d85f3ab25"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.466277 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "b13e03a7-b033-441d-9f0b-2c6d85f3ab25" (UID: "b13e03a7-b033-441d-9f0b-2c6d85f3ab25"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.466552 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-service-ca" (OuterVolumeSpecName: "service-ca") pod "b13e03a7-b033-441d-9f0b-2c6d85f3ab25" (UID: "b13e03a7-b033-441d-9f0b-2c6d85f3ab25"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.471776 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-kube-api-access-rc9lq" (OuterVolumeSpecName: "kube-api-access-rc9lq") pod "b13e03a7-b033-441d-9f0b-2c6d85f3ab25" (UID: "b13e03a7-b033-441d-9f0b-2c6d85f3ab25"). InnerVolumeSpecName "kube-api-access-rc9lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.472139 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "b13e03a7-b033-441d-9f0b-2c6d85f3ab25" (UID: "b13e03a7-b033-441d-9f0b-2c6d85f3ab25"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.472577 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "b13e03a7-b033-441d-9f0b-2c6d85f3ab25" (UID: "b13e03a7-b033-441d-9f0b-2c6d85f3ab25"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.567766 5035 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.567815 5035 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.567824 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc9lq\" (UniqueName: \"kubernetes.io/projected/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-kube-api-access-rc9lq\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.567834 5035 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.567842 5035 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.567850 5035 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:47 crc kubenswrapper[5035]: I1002 09:39:47.567858 5035 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b13e03a7-b033-441d-9f0b-2c6d85f3ab25-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.044954 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-b7fjd_b13e03a7-b033-441d-9f0b-2c6d85f3ab25/console/0.log" Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.044998 5035 generic.go:334] "Generic (PLEG): container finished" podID="b13e03a7-b033-441d-9f0b-2c6d85f3ab25" containerID="fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f" exitCode=2 Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.045033 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b7fjd" event={"ID":"b13e03a7-b033-441d-9f0b-2c6d85f3ab25","Type":"ContainerDied","Data":"fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f"} Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.045067 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-b7fjd" event={"ID":"b13e03a7-b033-441d-9f0b-2c6d85f3ab25","Type":"ContainerDied","Data":"d0c1778af46655c95daeefaff5f8bc785a6a8f43d85a9b1e598c026786cb02b8"} Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.045074 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-b7fjd" Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.045086 5035 scope.go:117] "RemoveContainer" containerID="fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f" Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.069117 5035 scope.go:117] "RemoveContainer" containerID="fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f" Oct 02 09:39:48 crc kubenswrapper[5035]: E1002 09:39:48.071393 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f\": container with ID starting with fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f not found: ID does not exist" containerID="fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f" Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.071574 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f"} err="failed to get container status \"fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f\": rpc error: code = NotFound desc = could not find container \"fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f\": container with ID starting with fa6f279494ab206537840d37beb1aecc4cd7a7bbb7af4a5d70ef6a63dce2d23f not found: ID does not exist" Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.074392 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-b7fjd"] Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.078483 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-b7fjd"] Oct 02 09:39:48 crc kubenswrapper[5035]: I1002 09:39:48.180807 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b13e03a7-b033-441d-9f0b-2c6d85f3ab25" path="/var/lib/kubelet/pods/b13e03a7-b033-441d-9f0b-2c6d85f3ab25/volumes" Oct 02 09:39:49 crc kubenswrapper[5035]: I1002 09:39:49.055722 5035 generic.go:334] "Generic (PLEG): container finished" podID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerID="ec47850eab11d74bc3e3dee1156bc8cd11052fc1d9f390c4c99651f9d4c306f6" exitCode=0 Oct 02 09:39:49 crc kubenswrapper[5035]: I1002 09:39:49.055799 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" event={"ID":"c643d696-0eaa-4495-acfe-6035dd231a0c","Type":"ContainerDied","Data":"ec47850eab11d74bc3e3dee1156bc8cd11052fc1d9f390c4c99651f9d4c306f6"} Oct 02 09:39:50 crc kubenswrapper[5035]: I1002 09:39:50.063553 5035 generic.go:334] "Generic (PLEG): container finished" podID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerID="a934685f41290190254c86ec7c4757be7bf76f3afc48b8aee12f12087998b361" exitCode=0 Oct 02 09:39:50 crc kubenswrapper[5035]: I1002 09:39:50.063603 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" event={"ID":"c643d696-0eaa-4495-acfe-6035dd231a0c","Type":"ContainerDied","Data":"a934685f41290190254c86ec7c4757be7bf76f3afc48b8aee12f12087998b361"} Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.325605 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.420804 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-bundle\") pod \"c643d696-0eaa-4495-acfe-6035dd231a0c\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.420928 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-util\") pod \"c643d696-0eaa-4495-acfe-6035dd231a0c\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.420976 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5jcp\" (UniqueName: \"kubernetes.io/projected/c643d696-0eaa-4495-acfe-6035dd231a0c-kube-api-access-k5jcp\") pod \"c643d696-0eaa-4495-acfe-6035dd231a0c\" (UID: \"c643d696-0eaa-4495-acfe-6035dd231a0c\") " Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.422467 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-bundle" (OuterVolumeSpecName: "bundle") pod "c643d696-0eaa-4495-acfe-6035dd231a0c" (UID: "c643d696-0eaa-4495-acfe-6035dd231a0c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.426069 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c643d696-0eaa-4495-acfe-6035dd231a0c-kube-api-access-k5jcp" (OuterVolumeSpecName: "kube-api-access-k5jcp") pod "c643d696-0eaa-4495-acfe-6035dd231a0c" (UID: "c643d696-0eaa-4495-acfe-6035dd231a0c"). InnerVolumeSpecName "kube-api-access-k5jcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.454386 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-util" (OuterVolumeSpecName: "util") pod "c643d696-0eaa-4495-acfe-6035dd231a0c" (UID: "c643d696-0eaa-4495-acfe-6035dd231a0c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.522360 5035 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.522395 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5jcp\" (UniqueName: \"kubernetes.io/projected/c643d696-0eaa-4495-acfe-6035dd231a0c-kube-api-access-k5jcp\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:51 crc kubenswrapper[5035]: I1002 09:39:51.522405 5035 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c643d696-0eaa-4495-acfe-6035dd231a0c-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:39:52 crc kubenswrapper[5035]: I1002 09:39:52.080020 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" event={"ID":"c643d696-0eaa-4495-acfe-6035dd231a0c","Type":"ContainerDied","Data":"079135d71be8a0ad5d6b2a933ddf263bb41f9183bc4220c28dafe24b1e0d10f3"} Oct 02 09:39:52 crc kubenswrapper[5035]: I1002 09:39:52.080514 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="079135d71be8a0ad5d6b2a933ddf263bb41f9183bc4220c28dafe24b1e0d10f3" Oct 02 09:39:52 crc kubenswrapper[5035]: I1002 09:39:52.080627 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48" Oct 02 09:39:55 crc kubenswrapper[5035]: I1002 09:39:55.538237 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:39:55 crc kubenswrapper[5035]: I1002 09:39:55.539293 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.859606 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9"] Oct 02 09:40:01 crc kubenswrapper[5035]: E1002 09:40:01.860339 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerName="util" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.860355 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerName="util" Oct 02 09:40:01 crc kubenswrapper[5035]: E1002 09:40:01.860371 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b13e03a7-b033-441d-9f0b-2c6d85f3ab25" containerName="console" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.860379 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b13e03a7-b033-441d-9f0b-2c6d85f3ab25" containerName="console" Oct 02 09:40:01 crc kubenswrapper[5035]: E1002 09:40:01.860391 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerName="pull" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.860399 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerName="pull" Oct 02 09:40:01 crc kubenswrapper[5035]: E1002 09:40:01.860414 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerName="extract" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.860422 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerName="extract" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.860543 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b13e03a7-b033-441d-9f0b-2c6d85f3ab25" containerName="console" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.860557 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c643d696-0eaa-4495-acfe-6035dd231a0c" containerName="extract" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.860913 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.863809 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.864001 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.864259 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zkv4s" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.864418 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.865239 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.872057 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9"] Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.956276 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9br9b\" (UniqueName: \"kubernetes.io/projected/e83c1515-25dc-45f1-9ff6-547382cce9b3-kube-api-access-9br9b\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.956344 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e83c1515-25dc-45f1-9ff6-547382cce9b3-webhook-cert\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:01 crc kubenswrapper[5035]: I1002 09:40:01.956723 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e83c1515-25dc-45f1-9ff6-547382cce9b3-apiservice-cert\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.058148 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e83c1515-25dc-45f1-9ff6-547382cce9b3-apiservice-cert\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.058238 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9br9b\" (UniqueName: \"kubernetes.io/projected/e83c1515-25dc-45f1-9ff6-547382cce9b3-kube-api-access-9br9b\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.058290 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e83c1515-25dc-45f1-9ff6-547382cce9b3-webhook-cert\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.068355 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e83c1515-25dc-45f1-9ff6-547382cce9b3-webhook-cert\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.069442 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e83c1515-25dc-45f1-9ff6-547382cce9b3-apiservice-cert\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.085701 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58"] Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.086478 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.087596 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9br9b\" (UniqueName: \"kubernetes.io/projected/e83c1515-25dc-45f1-9ff6-547382cce9b3-kube-api-access-9br9b\") pod \"metallb-operator-controller-manager-b9b6fcb4f-9qjh9\" (UID: \"e83c1515-25dc-45f1-9ff6-547382cce9b3\") " pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.088184 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.088280 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.088948 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-znpkw" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.104739 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58"] Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.159075 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b52d7\" (UniqueName: \"kubernetes.io/projected/cdbb31cd-2a07-479e-aebe-55a862352dc9-kube-api-access-b52d7\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.159151 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cdbb31cd-2a07-479e-aebe-55a862352dc9-webhook-cert\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.159182 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cdbb31cd-2a07-479e-aebe-55a862352dc9-apiservice-cert\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.226520 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.260225 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b52d7\" (UniqueName: \"kubernetes.io/projected/cdbb31cd-2a07-479e-aebe-55a862352dc9-kube-api-access-b52d7\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.260304 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cdbb31cd-2a07-479e-aebe-55a862352dc9-webhook-cert\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.260341 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cdbb31cd-2a07-479e-aebe-55a862352dc9-apiservice-cert\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.265289 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cdbb31cd-2a07-479e-aebe-55a862352dc9-apiservice-cert\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.266411 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cdbb31cd-2a07-479e-aebe-55a862352dc9-webhook-cert\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.276881 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b52d7\" (UniqueName: \"kubernetes.io/projected/cdbb31cd-2a07-479e-aebe-55a862352dc9-kube-api-access-b52d7\") pod \"metallb-operator-webhook-server-7cb8bcccbf-94b58\" (UID: \"cdbb31cd-2a07-479e-aebe-55a862352dc9\") " pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.440585 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.471280 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9"] Oct 02 09:40:02 crc kubenswrapper[5035]: I1002 09:40:02.690116 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58"] Oct 02 09:40:02 crc kubenswrapper[5035]: W1002 09:40:02.702936 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdbb31cd_2a07_479e_aebe_55a862352dc9.slice/crio-e370a409b035ac64f04aabbb4cb57bbf1537f2bac12fc5d6845ffcde165f0f5e WatchSource:0}: Error finding container e370a409b035ac64f04aabbb4cb57bbf1537f2bac12fc5d6845ffcde165f0f5e: Status 404 returned error can't find the container with id e370a409b035ac64f04aabbb4cb57bbf1537f2bac12fc5d6845ffcde165f0f5e Oct 02 09:40:03 crc kubenswrapper[5035]: I1002 09:40:03.153110 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" event={"ID":"cdbb31cd-2a07-479e-aebe-55a862352dc9","Type":"ContainerStarted","Data":"e370a409b035ac64f04aabbb4cb57bbf1537f2bac12fc5d6845ffcde165f0f5e"} Oct 02 09:40:03 crc kubenswrapper[5035]: I1002 09:40:03.154311 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" event={"ID":"e83c1515-25dc-45f1-9ff6-547382cce9b3","Type":"ContainerStarted","Data":"7ecd4c73ee2cbe4c5d42051afac47213b3ee0820193fc0740f61a5ac3e52e804"} Oct 02 09:40:08 crc kubenswrapper[5035]: I1002 09:40:08.182486 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" event={"ID":"e83c1515-25dc-45f1-9ff6-547382cce9b3","Type":"ContainerStarted","Data":"12781f1f90def821f5897c8ddb6d2431fe3f5ecb11a02809720313f6e4ae76a0"} Oct 02 09:40:08 crc kubenswrapper[5035]: I1002 09:40:08.184056 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:08 crc kubenswrapper[5035]: I1002 09:40:08.220098 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" podStartSLOduration=1.859296241 podStartE2EDuration="7.220078826s" podCreationTimestamp="2025-10-02 09:40:01 +0000 UTC" firstStartedPulling="2025-10-02 09:40:02.483699436 +0000 UTC m=+767.840043461" lastFinishedPulling="2025-10-02 09:40:07.844482021 +0000 UTC m=+773.200826046" observedRunningTime="2025-10-02 09:40:08.216228984 +0000 UTC m=+773.572573019" watchObservedRunningTime="2025-10-02 09:40:08.220078826 +0000 UTC m=+773.576422851" Oct 02 09:40:10 crc kubenswrapper[5035]: I1002 09:40:10.195106 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" event={"ID":"cdbb31cd-2a07-479e-aebe-55a862352dc9","Type":"ContainerStarted","Data":"8d97bc1f23ab28f9362c15330cef64cec3fc8a01e1d65dbd7156e62b5586affb"} Oct 02 09:40:10 crc kubenswrapper[5035]: I1002 09:40:10.196548 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.052935 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" podStartSLOduration=3.637078124 podStartE2EDuration="10.052914596s" podCreationTimestamp="2025-10-02 09:40:02 +0000 UTC" firstStartedPulling="2025-10-02 09:40:02.706186913 +0000 UTC m=+768.062530938" lastFinishedPulling="2025-10-02 09:40:09.122023385 +0000 UTC m=+774.478367410" observedRunningTime="2025-10-02 09:40:10.219358134 +0000 UTC m=+775.575702149" watchObservedRunningTime="2025-10-02 09:40:12.052914596 +0000 UTC m=+777.409258641" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.055442 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qb4qh"] Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.055697 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" podUID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" containerName="controller-manager" containerID="cri-o://7604f7d6bcf6ef92f6d526c19948e01af44108975e2a68dea6755a670d21a40e" gracePeriod=30 Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.153680 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6"] Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.153883 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" podUID="da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" containerName="route-controller-manager" containerID="cri-o://0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689" gracePeriod=30 Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.212993 5035 generic.go:334] "Generic (PLEG): container finished" podID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" containerID="7604f7d6bcf6ef92f6d526c19948e01af44108975e2a68dea6755a670d21a40e" exitCode=0 Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.213098 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" event={"ID":"8cabf8d0-11ee-4cb8-afb9-8dceb1461653","Type":"ContainerDied","Data":"7604f7d6bcf6ef92f6d526c19948e01af44108975e2a68dea6755a670d21a40e"} Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.482107 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.540450 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608272 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-client-ca\") pod \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608318 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-serving-cert\") pod \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608348 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdth5\" (UniqueName: \"kubernetes.io/projected/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-kube-api-access-xdth5\") pod \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608380 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-proxy-ca-bundles\") pod \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608410 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-config\") pod \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608427 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxgbv\" (UniqueName: \"kubernetes.io/projected/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-kube-api-access-pxgbv\") pod \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608458 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-serving-cert\") pod \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608507 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-config\") pod \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\" (UID: \"8cabf8d0-11ee-4cb8-afb9-8dceb1461653\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.608542 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-client-ca\") pod \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\" (UID: \"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae\") " Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.609166 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-config" (OuterVolumeSpecName: "config") pod "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" (UID: "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.609167 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8cabf8d0-11ee-4cb8-afb9-8dceb1461653" (UID: "8cabf8d0-11ee-4cb8-afb9-8dceb1461653"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.609198 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-config" (OuterVolumeSpecName: "config") pod "8cabf8d0-11ee-4cb8-afb9-8dceb1461653" (UID: "8cabf8d0-11ee-4cb8-afb9-8dceb1461653"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.609215 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-client-ca" (OuterVolumeSpecName: "client-ca") pod "8cabf8d0-11ee-4cb8-afb9-8dceb1461653" (UID: "8cabf8d0-11ee-4cb8-afb9-8dceb1461653"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.609188 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-client-ca" (OuterVolumeSpecName: "client-ca") pod "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" (UID: "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.613607 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-kube-api-access-xdth5" (OuterVolumeSpecName: "kube-api-access-xdth5") pod "8cabf8d0-11ee-4cb8-afb9-8dceb1461653" (UID: "8cabf8d0-11ee-4cb8-afb9-8dceb1461653"). InnerVolumeSpecName "kube-api-access-xdth5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.613776 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-kube-api-access-pxgbv" (OuterVolumeSpecName: "kube-api-access-pxgbv") pod "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" (UID: "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae"). InnerVolumeSpecName "kube-api-access-pxgbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.613833 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" (UID: "da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.614139 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cabf8d0-11ee-4cb8-afb9-8dceb1461653" (UID: "8cabf8d0-11ee-4cb8-afb9-8dceb1461653"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710280 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710315 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxgbv\" (UniqueName: \"kubernetes.io/projected/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-kube-api-access-pxgbv\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710334 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710346 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710359 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710369 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710379 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710390 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdth5\" (UniqueName: \"kubernetes.io/projected/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-kube-api-access-xdth5\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:12 crc kubenswrapper[5035]: I1002 09:40:12.710400 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8cabf8d0-11ee-4cb8-afb9-8dceb1461653-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.118899 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cffdfc965-zzwr9"] Oct 02 09:40:13 crc kubenswrapper[5035]: E1002 09:40:13.119096 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" containerName="route-controller-manager" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.119106 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" containerName="route-controller-manager" Oct 02 09:40:13 crc kubenswrapper[5035]: E1002 09:40:13.119117 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" containerName="controller-manager" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.119123 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" containerName="controller-manager" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.119241 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" containerName="route-controller-manager" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.119252 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" containerName="controller-manager" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.119628 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.129952 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cffdfc965-zzwr9"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.216377 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-proxy-ca-bundles\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.216808 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-client-ca\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.216852 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-config\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.216891 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33cc1c09-3339-40fe-923f-d62b4d0d4e86-serving-cert\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.216919 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m2cw\" (UniqueName: \"kubernetes.io/projected/33cc1c09-3339-40fe-923f-d62b4d0d4e86-kube-api-access-7m2cw\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.221146 5035 generic.go:334] "Generic (PLEG): container finished" podID="da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" containerID="0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689" exitCode=0 Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.221190 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" event={"ID":"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae","Type":"ContainerDied","Data":"0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689"} Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.221203 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.221369 5035 scope.go:117] "RemoveContainer" containerID="0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.221508 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6" event={"ID":"da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae","Type":"ContainerDied","Data":"725be3898fb7302ec7b03ab2a1f5b5f008b3a32b2caca5524da4ae9149ddaa96"} Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.222587 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" event={"ID":"8cabf8d0-11ee-4cb8-afb9-8dceb1461653","Type":"ContainerDied","Data":"887ca1a3aa61f8ddee40e8e68692e215633820c4b20804b3f2af441ed396cb60"} Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.222653 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qb4qh" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.242555 5035 scope.go:117] "RemoveContainer" containerID="0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689" Oct 02 09:40:13 crc kubenswrapper[5035]: E1002 09:40:13.244341 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689\": container with ID starting with 0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689 not found: ID does not exist" containerID="0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.244373 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689"} err="failed to get container status \"0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689\": rpc error: code = NotFound desc = could not find container \"0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689\": container with ID starting with 0946e1073b4d8d66af00704517b89337a383e574e351f0850b96e370c9801689 not found: ID does not exist" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.244399 5035 scope.go:117] "RemoveContainer" containerID="7604f7d6bcf6ef92f6d526c19948e01af44108975e2a68dea6755a670d21a40e" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.257263 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qb4qh"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.270904 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qb4qh"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.280898 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.283651 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-ldpj6"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.318016 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-proxy-ca-bundles\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.318252 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-client-ca\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.318325 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-config\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.318432 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33cc1c09-3339-40fe-923f-d62b4d0d4e86-serving-cert\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.318510 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m2cw\" (UniqueName: \"kubernetes.io/projected/33cc1c09-3339-40fe-923f-d62b4d0d4e86-kube-api-access-7m2cw\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.319192 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-client-ca\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.319223 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-proxy-ca-bundles\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.320096 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-config\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.321906 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33cc1c09-3339-40fe-923f-d62b4d0d4e86-serving-cert\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.333212 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m2cw\" (UniqueName: \"kubernetes.io/projected/33cc1c09-3339-40fe-923f-d62b4d0d4e86-kube-api-access-7m2cw\") pod \"controller-manager-7cffdfc965-zzwr9\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.432526 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.762211 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cffdfc965-zzwr9"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.792863 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.793753 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.796907 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.797071 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.797672 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.797918 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.798153 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.803405 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2"] Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.803778 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.925434 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-serving-cert\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.925554 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-config\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.925577 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6chv\" (UniqueName: \"kubernetes.io/projected/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-kube-api-access-k6chv\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.925610 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-client-ca\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:13 crc kubenswrapper[5035]: I1002 09:40:13.957239 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cffdfc965-zzwr9"] Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.026768 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-config\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.026811 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6chv\" (UniqueName: \"kubernetes.io/projected/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-kube-api-access-k6chv\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.026858 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-client-ca\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.026903 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-serving-cert\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.028262 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-client-ca\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.028449 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-config\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.032834 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-serving-cert\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.045821 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6chv\" (UniqueName: \"kubernetes.io/projected/09b1ad61-eea1-46c5-afb2-78ed3bb914a6-kube-api-access-k6chv\") pod \"route-controller-manager-655fff85ff-b5dt2\" (UID: \"09b1ad61-eea1-46c5-afb2-78ed3bb914a6\") " pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.114468 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.173912 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cabf8d0-11ee-4cb8-afb9-8dceb1461653" path="/var/lib/kubelet/pods/8cabf8d0-11ee-4cb8-afb9-8dceb1461653/volumes" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.174802 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae" path="/var/lib/kubelet/pods/da5b4cd7-495f-43c5-b0fd-e96fce9fa2ae/volumes" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.235273 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" event={"ID":"33cc1c09-3339-40fe-923f-d62b4d0d4e86","Type":"ContainerStarted","Data":"c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767"} Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.235322 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" event={"ID":"33cc1c09-3339-40fe-923f-d62b4d0d4e86","Type":"ContainerStarted","Data":"3e23c6c0049f69d067d8417f315448b83689cd01347032bfbd0ab272a6aecf72"} Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.235396 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" podUID="33cc1c09-3339-40fe-923f-d62b4d0d4e86" containerName="controller-manager" containerID="cri-o://c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767" gracePeriod=30 Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.235852 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.242043 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.259634 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" podStartSLOduration=2.259617633 podStartE2EDuration="2.259617633s" podCreationTimestamp="2025-10-02 09:40:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:40:14.257984806 +0000 UTC m=+779.614328831" watchObservedRunningTime="2025-10-02 09:40:14.259617633 +0000 UTC m=+779.615961658" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.523183 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2"] Oct 02 09:40:14 crc kubenswrapper[5035]: W1002 09:40:14.535604 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09b1ad61_eea1_46c5_afb2_78ed3bb914a6.slice/crio-b34bef805a2873b4f473e0b9f85277934ae5ab4d61ae906959e098e31184a88c WatchSource:0}: Error finding container b34bef805a2873b4f473e0b9f85277934ae5ab4d61ae906959e098e31184a88c: Status 404 returned error can't find the container with id b34bef805a2873b4f473e0b9f85277934ae5ab4d61ae906959e098e31184a88c Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.616551 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.737934 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-proxy-ca-bundles\") pod \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.738006 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m2cw\" (UniqueName: \"kubernetes.io/projected/33cc1c09-3339-40fe-923f-d62b4d0d4e86-kube-api-access-7m2cw\") pod \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.738056 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-client-ca\") pod \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.738090 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33cc1c09-3339-40fe-923f-d62b4d0d4e86-serving-cert\") pod \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.738129 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-config\") pod \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\" (UID: \"33cc1c09-3339-40fe-923f-d62b4d0d4e86\") " Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.739391 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-client-ca" (OuterVolumeSpecName: "client-ca") pod "33cc1c09-3339-40fe-923f-d62b4d0d4e86" (UID: "33cc1c09-3339-40fe-923f-d62b4d0d4e86"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.739855 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-config" (OuterVolumeSpecName: "config") pod "33cc1c09-3339-40fe-923f-d62b4d0d4e86" (UID: "33cc1c09-3339-40fe-923f-d62b4d0d4e86"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.740648 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "33cc1c09-3339-40fe-923f-d62b4d0d4e86" (UID: "33cc1c09-3339-40fe-923f-d62b4d0d4e86"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.743918 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33cc1c09-3339-40fe-923f-d62b4d0d4e86-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "33cc1c09-3339-40fe-923f-d62b4d0d4e86" (UID: "33cc1c09-3339-40fe-923f-d62b4d0d4e86"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.744302 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33cc1c09-3339-40fe-923f-d62b4d0d4e86-kube-api-access-7m2cw" (OuterVolumeSpecName: "kube-api-access-7m2cw") pod "33cc1c09-3339-40fe-923f-d62b4d0d4e86" (UID: "33cc1c09-3339-40fe-923f-d62b4d0d4e86"). InnerVolumeSpecName "kube-api-access-7m2cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.840297 5035 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.840375 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m2cw\" (UniqueName: \"kubernetes.io/projected/33cc1c09-3339-40fe-923f-d62b4d0d4e86-kube-api-access-7m2cw\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.840401 5035 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.840419 5035 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33cc1c09-3339-40fe-923f-d62b4d0d4e86-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:14 crc kubenswrapper[5035]: I1002 09:40:14.840442 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33cc1c09-3339-40fe-923f-d62b4d0d4e86-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.121313 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77c59f8885-kj49h"] Oct 02 09:40:15 crc kubenswrapper[5035]: E1002 09:40:15.121684 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33cc1c09-3339-40fe-923f-d62b4d0d4e86" containerName="controller-manager" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.121709 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="33cc1c09-3339-40fe-923f-d62b4d0d4e86" containerName="controller-manager" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.121919 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="33cc1c09-3339-40fe-923f-d62b4d0d4e86" containerName="controller-manager" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.124277 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.142276 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77c59f8885-kj49h"] Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.242649 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" event={"ID":"09b1ad61-eea1-46c5-afb2-78ed3bb914a6","Type":"ContainerStarted","Data":"1fee32571dfb84b4f8801d6b139fdaca9b77d236de3f7b671a258d4c03d9ac94"} Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.242701 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" event={"ID":"09b1ad61-eea1-46c5-afb2-78ed3bb914a6","Type":"ContainerStarted","Data":"b34bef805a2873b4f473e0b9f85277934ae5ab4d61ae906959e098e31184a88c"} Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.244381 5035 generic.go:334] "Generic (PLEG): container finished" podID="33cc1c09-3339-40fe-923f-d62b4d0d4e86" containerID="c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767" exitCode=0 Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.244420 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" event={"ID":"33cc1c09-3339-40fe-923f-d62b4d0d4e86","Type":"ContainerDied","Data":"c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767"} Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.244444 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" event={"ID":"33cc1c09-3339-40fe-923f-d62b4d0d4e86","Type":"ContainerDied","Data":"3e23c6c0049f69d067d8417f315448b83689cd01347032bfbd0ab272a6aecf72"} Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.244451 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cffdfc965-zzwr9" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.244464 5035 scope.go:117] "RemoveContainer" containerID="c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.247975 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-config\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.248346 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-proxy-ca-bundles\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.248420 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9d48\" (UniqueName: \"kubernetes.io/projected/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-kube-api-access-w9d48\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.248463 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-serving-cert\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.248490 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-client-ca\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.260248 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" podStartSLOduration=2.260231036 podStartE2EDuration="2.260231036s" podCreationTimestamp="2025-10-02 09:40:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:40:15.258088053 +0000 UTC m=+780.614432098" watchObservedRunningTime="2025-10-02 09:40:15.260231036 +0000 UTC m=+780.616575061" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.262908 5035 scope.go:117] "RemoveContainer" containerID="c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767" Oct 02 09:40:15 crc kubenswrapper[5035]: E1002 09:40:15.264285 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767\": container with ID starting with c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767 not found: ID does not exist" containerID="c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.264337 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767"} err="failed to get container status \"c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767\": rpc error: code = NotFound desc = could not find container \"c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767\": container with ID starting with c361d3896b8188f478f2fcb862f67491c0907b69f42277e1de582fae265e2767 not found: ID does not exist" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.277182 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cffdfc965-zzwr9"] Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.279969 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7cffdfc965-zzwr9"] Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.349264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-config\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.349332 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-proxy-ca-bundles\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.349441 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9d48\" (UniqueName: \"kubernetes.io/projected/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-kube-api-access-w9d48\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.349554 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-serving-cert\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.349606 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-client-ca\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.350844 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-proxy-ca-bundles\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.350922 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-client-ca\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.351281 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-config\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.362316 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-serving-cert\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.365973 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9d48\" (UniqueName: \"kubernetes.io/projected/2ed7c1fc-4936-4de0-a1aa-26d1e0b10642-kube-api-access-w9d48\") pod \"controller-manager-77c59f8885-kj49h\" (UID: \"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642\") " pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.439981 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:15 crc kubenswrapper[5035]: I1002 09:40:15.706910 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77c59f8885-kj49h"] Oct 02 09:40:15 crc kubenswrapper[5035]: W1002 09:40:15.717311 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ed7c1fc_4936_4de0_a1aa_26d1e0b10642.slice/crio-896df29959cb16bf0b9bdced0cfe9438ea98b9d332cc324dabe512eab20dd7d6 WatchSource:0}: Error finding container 896df29959cb16bf0b9bdced0cfe9438ea98b9d332cc324dabe512eab20dd7d6: Status 404 returned error can't find the container with id 896df29959cb16bf0b9bdced0cfe9438ea98b9d332cc324dabe512eab20dd7d6 Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.170987 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33cc1c09-3339-40fe-923f-d62b4d0d4e86" path="/var/lib/kubelet/pods/33cc1c09-3339-40fe-923f-d62b4d0d4e86/volumes" Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.250028 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" event={"ID":"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642","Type":"ContainerStarted","Data":"cc8aa53e54286d823a71c3a7884d1ad03b364b9df1ab2ea141e0abbcd0acb10e"} Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.250388 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" event={"ID":"2ed7c1fc-4936-4de0-a1aa-26d1e0b10642","Type":"ContainerStarted","Data":"896df29959cb16bf0b9bdced0cfe9438ea98b9d332cc324dabe512eab20dd7d6"} Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.250407 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.251577 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.255011 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.255878 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-655fff85ff-b5dt2" Oct 02 09:40:16 crc kubenswrapper[5035]: I1002 09:40:16.273606 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77c59f8885-kj49h" podStartSLOduration=3.273589698 podStartE2EDuration="3.273589698s" podCreationTimestamp="2025-10-02 09:40:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:40:16.271699593 +0000 UTC m=+781.628043618" watchObservedRunningTime="2025-10-02 09:40:16.273589698 +0000 UTC m=+781.629933723" Oct 02 09:40:18 crc kubenswrapper[5035]: I1002 09:40:18.354381 5035 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 09:40:22 crc kubenswrapper[5035]: I1002 09:40:22.447323 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7cb8bcccbf-94b58" Oct 02 09:40:25 crc kubenswrapper[5035]: I1002 09:40:25.538218 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:40:25 crc kubenswrapper[5035]: I1002 09:40:25.538574 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:40:32 crc kubenswrapper[5035]: I1002 09:40:32.949908 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hj8c8"] Oct 02 09:40:32 crc kubenswrapper[5035]: I1002 09:40:32.953217 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:32 crc kubenswrapper[5035]: I1002 09:40:32.958604 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hj8c8"] Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.076123 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-catalog-content\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.076214 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-utilities\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.076286 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk29r\" (UniqueName: \"kubernetes.io/projected/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-kube-api-access-lk29r\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.177981 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-catalog-content\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.178031 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-utilities\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.178050 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk29r\" (UniqueName: \"kubernetes.io/projected/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-kube-api-access-lk29r\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.178791 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-catalog-content\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.179003 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-utilities\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.207407 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk29r\" (UniqueName: \"kubernetes.io/projected/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-kube-api-access-lk29r\") pod \"redhat-operators-hj8c8\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.275016 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:33 crc kubenswrapper[5035]: I1002 09:40:33.661421 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hj8c8"] Oct 02 09:40:33 crc kubenswrapper[5035]: W1002 09:40:33.670694 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c6b787b_3b9d_49d4_a51f_cd4a98f348e1.slice/crio-ca3d6226f627cf1c4a206950d6756cfb9b3c2b2c31274c2a3bc19cbe24f9bcb3 WatchSource:0}: Error finding container ca3d6226f627cf1c4a206950d6756cfb9b3c2b2c31274c2a3bc19cbe24f9bcb3: Status 404 returned error can't find the container with id ca3d6226f627cf1c4a206950d6756cfb9b3c2b2c31274c2a3bc19cbe24f9bcb3 Oct 02 09:40:34 crc kubenswrapper[5035]: I1002 09:40:34.353170 5035 generic.go:334] "Generic (PLEG): container finished" podID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerID="2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3" exitCode=0 Oct 02 09:40:34 crc kubenswrapper[5035]: I1002 09:40:34.353213 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8c8" event={"ID":"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1","Type":"ContainerDied","Data":"2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3"} Oct 02 09:40:34 crc kubenswrapper[5035]: I1002 09:40:34.353242 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8c8" event={"ID":"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1","Type":"ContainerStarted","Data":"ca3d6226f627cf1c4a206950d6756cfb9b3c2b2c31274c2a3bc19cbe24f9bcb3"} Oct 02 09:40:35 crc kubenswrapper[5035]: I1002 09:40:35.360201 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8c8" event={"ID":"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1","Type":"ContainerStarted","Data":"3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33"} Oct 02 09:40:36 crc kubenswrapper[5035]: I1002 09:40:36.367218 5035 generic.go:334] "Generic (PLEG): container finished" podID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerID="3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33" exitCode=0 Oct 02 09:40:36 crc kubenswrapper[5035]: I1002 09:40:36.367331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8c8" event={"ID":"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1","Type":"ContainerDied","Data":"3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33"} Oct 02 09:40:37 crc kubenswrapper[5035]: I1002 09:40:37.380700 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8c8" event={"ID":"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1","Type":"ContainerStarted","Data":"9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60"} Oct 02 09:40:42 crc kubenswrapper[5035]: I1002 09:40:42.229621 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-b9b6fcb4f-9qjh9" Oct 02 09:40:42 crc kubenswrapper[5035]: I1002 09:40:42.250386 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hj8c8" podStartSLOduration=7.700428008 podStartE2EDuration="10.250365779s" podCreationTimestamp="2025-10-02 09:40:32 +0000 UTC" firstStartedPulling="2025-10-02 09:40:34.355014052 +0000 UTC m=+799.711358077" lastFinishedPulling="2025-10-02 09:40:36.904951823 +0000 UTC m=+802.261295848" observedRunningTime="2025-10-02 09:40:37.404389422 +0000 UTC m=+802.760733457" watchObservedRunningTime="2025-10-02 09:40:42.250365779 +0000 UTC m=+807.606709804" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.016934 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.017715 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.021032 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-576rd"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.023118 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.023365 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-w5m9z" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.024665 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.024692 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.048566 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.054985 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.106388 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-kn2pw"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.107215 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.109166 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m27xz\" (UniqueName: \"kubernetes.io/projected/3b2dae08-474c-4f53-802a-35133f5f5119-kube-api-access-m27xz\") pod \"frr-k8s-webhook-server-64bf5d555-9qrxg\" (UID: \"3b2dae08-474c-4f53-802a-35133f5f5119\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.109215 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b2dae08-474c-4f53-802a-35133f5f5119-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9qrxg\" (UID: \"3b2dae08-474c-4f53-802a-35133f5f5119\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.110008 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.110238 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.110368 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.110515 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-bprgc" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.118433 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-sms6d"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.119385 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.122773 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.132654 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-sms6d"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.210807 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-reloader\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.210883 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.210969 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b2dae08-474c-4f53-802a-35133f5f5119-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9qrxg\" (UID: \"3b2dae08-474c-4f53-802a-35133f5f5119\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.210996 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4536e046-502a-4344-87d4-d05e53f3a73b-frr-startup\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211014 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-metrics\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211030 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4536e046-502a-4344-87d4-d05e53f3a73b-metrics-certs\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211063 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-metallb-excludel2\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211085 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-metrics-certs\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211106 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebcd114a-13c7-4c49-a893-7016d73a6e0c-cert\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211127 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-frr-conf\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211141 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgrtm\" (UniqueName: \"kubernetes.io/projected/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-kube-api-access-xgrtm\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211164 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m27xz\" (UniqueName: \"kubernetes.io/projected/3b2dae08-474c-4f53-802a-35133f5f5119-kube-api-access-m27xz\") pod \"frr-k8s-webhook-server-64bf5d555-9qrxg\" (UID: \"3b2dae08-474c-4f53-802a-35133f5f5119\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211183 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-frr-sockets\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211200 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebcd114a-13c7-4c49-a893-7016d73a6e0c-metrics-certs\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211216 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nzmw\" (UniqueName: \"kubernetes.io/projected/ebcd114a-13c7-4c49-a893-7016d73a6e0c-kube-api-access-4nzmw\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.211233 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wwv6\" (UniqueName: \"kubernetes.io/projected/4536e046-502a-4344-87d4-d05e53f3a73b-kube-api-access-9wwv6\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: E1002 09:40:43.211359 5035 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 02 09:40:43 crc kubenswrapper[5035]: E1002 09:40:43.211404 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3b2dae08-474c-4f53-802a-35133f5f5119-cert podName:3b2dae08-474c-4f53-802a-35133f5f5119 nodeName:}" failed. No retries permitted until 2025-10-02 09:40:43.711388299 +0000 UTC m=+809.067732324 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3b2dae08-474c-4f53-802a-35133f5f5119-cert") pod "frr-k8s-webhook-server-64bf5d555-9qrxg" (UID: "3b2dae08-474c-4f53-802a-35133f5f5119") : secret "frr-k8s-webhook-server-cert" not found Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.247667 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m27xz\" (UniqueName: \"kubernetes.io/projected/3b2dae08-474c-4f53-802a-35133f5f5119-kube-api-access-m27xz\") pod \"frr-k8s-webhook-server-64bf5d555-9qrxg\" (UID: \"3b2dae08-474c-4f53-802a-35133f5f5119\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.275930 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.275989 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313064 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebcd114a-13c7-4c49-a893-7016d73a6e0c-metrics-certs\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313106 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nzmw\" (UniqueName: \"kubernetes.io/projected/ebcd114a-13c7-4c49-a893-7016d73a6e0c-kube-api-access-4nzmw\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313124 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-frr-sockets\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313142 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wwv6\" (UniqueName: \"kubernetes.io/projected/4536e046-502a-4344-87d4-d05e53f3a73b-kube-api-access-9wwv6\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313160 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-reloader\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313175 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313227 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4536e046-502a-4344-87d4-d05e53f3a73b-frr-startup\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313251 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-metrics\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313269 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4536e046-502a-4344-87d4-d05e53f3a73b-metrics-certs\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313300 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-metallb-excludel2\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313321 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-metrics-certs\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313341 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebcd114a-13c7-4c49-a893-7016d73a6e0c-cert\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313358 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-frr-conf\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.313374 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgrtm\" (UniqueName: \"kubernetes.io/projected/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-kube-api-access-xgrtm\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: E1002 09:40:43.314024 5035 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 09:40:43 crc kubenswrapper[5035]: E1002 09:40:43.314118 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist podName:aa3805ad-fdc2-4c45-9c0e-6c0d933d8351 nodeName:}" failed. No retries permitted until 2025-10-02 09:40:43.814092639 +0000 UTC m=+809.170436664 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist") pod "speaker-kn2pw" (UID: "aa3805ad-fdc2-4c45-9c0e-6c0d933d8351") : secret "metallb-memberlist" not found Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.314113 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-metrics\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.314297 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-frr-conf\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.314324 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-frr-sockets\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.314704 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4536e046-502a-4344-87d4-d05e53f3a73b-reloader\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.314826 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4536e046-502a-4344-87d4-d05e53f3a73b-frr-startup\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.314888 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-metallb-excludel2\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.317839 5035 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.318080 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4536e046-502a-4344-87d4-d05e53f3a73b-metrics-certs\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.318078 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-metrics-certs\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.319488 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.324093 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ebcd114a-13c7-4c49-a893-7016d73a6e0c-metrics-certs\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.330082 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebcd114a-13c7-4c49-a893-7016d73a6e0c-cert\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.330214 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgrtm\" (UniqueName: \"kubernetes.io/projected/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-kube-api-access-xgrtm\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.335387 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nzmw\" (UniqueName: \"kubernetes.io/projected/ebcd114a-13c7-4c49-a893-7016d73a6e0c-kube-api-access-4nzmw\") pod \"controller-68d546b9d8-sms6d\" (UID: \"ebcd114a-13c7-4c49-a893-7016d73a6e0c\") " pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.339010 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wwv6\" (UniqueName: \"kubernetes.io/projected/4536e046-502a-4344-87d4-d05e53f3a73b-kube-api-access-9wwv6\") pod \"frr-k8s-576rd\" (UID: \"4536e046-502a-4344-87d4-d05e53f3a73b\") " pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.351925 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.432390 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.453717 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.549194 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hj8c8"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.718109 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b2dae08-474c-4f53-802a-35133f5f5119-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9qrxg\" (UID: \"3b2dae08-474c-4f53-802a-35133f5f5119\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.722013 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3b2dae08-474c-4f53-802a-35133f5f5119-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9qrxg\" (UID: \"3b2dae08-474c-4f53-802a-35133f5f5119\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.819547 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:43 crc kubenswrapper[5035]: E1002 09:40:43.819810 5035 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 09:40:43 crc kubenswrapper[5035]: E1002 09:40:43.819904 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist podName:aa3805ad-fdc2-4c45-9c0e-6c0d933d8351 nodeName:}" failed. No retries permitted until 2025-10-02 09:40:44.819874334 +0000 UTC m=+810.176218399 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist") pod "speaker-kn2pw" (UID: "aa3805ad-fdc2-4c45-9c0e-6c0d933d8351") : secret "metallb-memberlist" not found Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.829637 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-sms6d"] Oct 02 09:40:43 crc kubenswrapper[5035]: I1002 09:40:43.935655 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.350331 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg"] Oct 02 09:40:44 crc kubenswrapper[5035]: W1002 09:40:44.359389 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b2dae08_474c_4f53_802a_35133f5f5119.slice/crio-1d0ccf2b98bf14f7b27626f6a8057dfad0dde13b9ecc3bd6f27fc0b784daf728 WatchSource:0}: Error finding container 1d0ccf2b98bf14f7b27626f6a8057dfad0dde13b9ecc3bd6f27fc0b784daf728: Status 404 returned error can't find the container with id 1d0ccf2b98bf14f7b27626f6a8057dfad0dde13b9ecc3bd6f27fc0b784daf728 Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.417748 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-sms6d" event={"ID":"ebcd114a-13c7-4c49-a893-7016d73a6e0c","Type":"ContainerStarted","Data":"47801e24d83f0a4d3666ebb70e9aa70d5fb1f30d9b878dfbcb4204a2b61f9433"} Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.417800 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-sms6d" event={"ID":"ebcd114a-13c7-4c49-a893-7016d73a6e0c","Type":"ContainerStarted","Data":"6bf79ce6e4e11c7534052c0533cd8f637a873e7f9706d1003e8caa1b995fe3fc"} Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.417818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-sms6d" event={"ID":"ebcd114a-13c7-4c49-a893-7016d73a6e0c","Type":"ContainerStarted","Data":"ef1816e59a428a9cc750e17967b3dddc9eb2af36ff77284684e4022aa9726641"} Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.417926 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.419214 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerStarted","Data":"0c2f6f68b2999f1b773d2702326571fa471fa7b560bdb019a901b936144ee1e6"} Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.421093 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" event={"ID":"3b2dae08-474c-4f53-802a-35133f5f5119","Type":"ContainerStarted","Data":"1d0ccf2b98bf14f7b27626f6a8057dfad0dde13b9ecc3bd6f27fc0b784daf728"} Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.833627 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.850365 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/aa3805ad-fdc2-4c45-9c0e-6c0d933d8351-memberlist\") pod \"speaker-kn2pw\" (UID: \"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351\") " pod="metallb-system/speaker-kn2pw" Oct 02 09:40:44 crc kubenswrapper[5035]: I1002 09:40:44.922221 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-kn2pw" Oct 02 09:40:44 crc kubenswrapper[5035]: W1002 09:40:44.940957 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa3805ad_fdc2_4c45_9c0e_6c0d933d8351.slice/crio-fa1d3897075a2101f55416059026b99f385257f2bf6b652a2ed4235bd2084749 WatchSource:0}: Error finding container fa1d3897075a2101f55416059026b99f385257f2bf6b652a2ed4235bd2084749: Status 404 returned error can't find the container with id fa1d3897075a2101f55416059026b99f385257f2bf6b652a2ed4235bd2084749 Oct 02 09:40:45 crc kubenswrapper[5035]: I1002 09:40:45.435003 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-kn2pw" event={"ID":"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351","Type":"ContainerStarted","Data":"b808a67ea38fe54799347a886806fd153fbfa7a788719932a3fcc4b76735d0b3"} Oct 02 09:40:45 crc kubenswrapper[5035]: I1002 09:40:45.435300 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-kn2pw" event={"ID":"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351","Type":"ContainerStarted","Data":"8f6654d5f5fb3a6cea0ee85816f00068603397ac5e4f86f92c9d89855fe49e04"} Oct 02 09:40:45 crc kubenswrapper[5035]: I1002 09:40:45.435314 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-kn2pw" event={"ID":"aa3805ad-fdc2-4c45-9c0e-6c0d933d8351","Type":"ContainerStarted","Data":"fa1d3897075a2101f55416059026b99f385257f2bf6b652a2ed4235bd2084749"} Oct 02 09:40:45 crc kubenswrapper[5035]: I1002 09:40:45.435608 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hj8c8" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="registry-server" containerID="cri-o://9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60" gracePeriod=2 Oct 02 09:40:45 crc kubenswrapper[5035]: I1002 09:40:45.435714 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-kn2pw" Oct 02 09:40:45 crc kubenswrapper[5035]: I1002 09:40:45.462947 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-sms6d" podStartSLOduration=2.46291432 podStartE2EDuration="2.46291432s" podCreationTimestamp="2025-10-02 09:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:40:44.43524774 +0000 UTC m=+809.791591765" watchObservedRunningTime="2025-10-02 09:40:45.46291432 +0000 UTC m=+810.819258345" Oct 02 09:40:45 crc kubenswrapper[5035]: I1002 09:40:45.465695 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-kn2pw" podStartSLOduration=2.465683581 podStartE2EDuration="2.465683581s" podCreationTimestamp="2025-10-02 09:40:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:40:45.455579486 +0000 UTC m=+810.811923531" watchObservedRunningTime="2025-10-02 09:40:45.465683581 +0000 UTC m=+810.822027606" Oct 02 09:40:45 crc kubenswrapper[5035]: E1002 09:40:45.604385 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c6b787b_3b9d_49d4_a51f_cd4a98f348e1.slice/crio-9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.016405 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.157915 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-catalog-content\") pod \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.158018 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk29r\" (UniqueName: \"kubernetes.io/projected/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-kube-api-access-lk29r\") pod \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.158092 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-utilities\") pod \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\" (UID: \"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1\") " Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.159066 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-utilities" (OuterVolumeSpecName: "utilities") pod "5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" (UID: "5c6b787b-3b9d-49d4-a51f-cd4a98f348e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.163268 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-kube-api-access-lk29r" (OuterVolumeSpecName: "kube-api-access-lk29r") pod "5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" (UID: "5c6b787b-3b9d-49d4-a51f-cd4a98f348e1"). InnerVolumeSpecName "kube-api-access-lk29r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.259424 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk29r\" (UniqueName: \"kubernetes.io/projected/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-kube-api-access-lk29r\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.259472 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.449414 5035 generic.go:334] "Generic (PLEG): container finished" podID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerID="9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60" exitCode=0 Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.449479 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hj8c8" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.449506 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8c8" event={"ID":"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1","Type":"ContainerDied","Data":"9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60"} Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.449564 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hj8c8" event={"ID":"5c6b787b-3b9d-49d4-a51f-cd4a98f348e1","Type":"ContainerDied","Data":"ca3d6226f627cf1c4a206950d6756cfb9b3c2b2c31274c2a3bc19cbe24f9bcb3"} Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.449587 5035 scope.go:117] "RemoveContainer" containerID="9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.478612 5035 scope.go:117] "RemoveContainer" containerID="3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.504710 5035 scope.go:117] "RemoveContainer" containerID="2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.547402 5035 scope.go:117] "RemoveContainer" containerID="9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60" Oct 02 09:40:46 crc kubenswrapper[5035]: E1002 09:40:46.547962 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60\": container with ID starting with 9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60 not found: ID does not exist" containerID="9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.547998 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60"} err="failed to get container status \"9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60\": rpc error: code = NotFound desc = could not find container \"9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60\": container with ID starting with 9821433c73bd9bcec5c0fb0b0045de03ed631a41c496ba1714cc8c60e51f1c60 not found: ID does not exist" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.548026 5035 scope.go:117] "RemoveContainer" containerID="3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33" Oct 02 09:40:46 crc kubenswrapper[5035]: E1002 09:40:46.553084 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33\": container with ID starting with 3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33 not found: ID does not exist" containerID="3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.553132 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33"} err="failed to get container status \"3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33\": rpc error: code = NotFound desc = could not find container \"3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33\": container with ID starting with 3786e55527301f038c41d9c906b5d1701e155be2c20d03a97a8d349c30279b33 not found: ID does not exist" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.553171 5035 scope.go:117] "RemoveContainer" containerID="2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3" Oct 02 09:40:46 crc kubenswrapper[5035]: E1002 09:40:46.556877 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3\": container with ID starting with 2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3 not found: ID does not exist" containerID="2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3" Oct 02 09:40:46 crc kubenswrapper[5035]: I1002 09:40:46.557120 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3"} err="failed to get container status \"2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3\": rpc error: code = NotFound desc = could not find container \"2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3\": container with ID starting with 2119827f4da4cc462d0a2f6223095b334643ad85a0c04386278db95e58a154a3 not found: ID does not exist" Oct 02 09:40:47 crc kubenswrapper[5035]: I1002 09:40:47.621791 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" (UID: "5c6b787b-3b9d-49d4-a51f-cd4a98f348e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:40:47 crc kubenswrapper[5035]: I1002 09:40:47.678570 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hj8c8"] Oct 02 09:40:47 crc kubenswrapper[5035]: I1002 09:40:47.679908 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:40:47 crc kubenswrapper[5035]: I1002 09:40:47.683237 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hj8c8"] Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.170650 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" path="/var/lib/kubelet/pods/5c6b787b-3b9d-49d4-a51f-cd4a98f348e1/volumes" Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.967646 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bn74v"] Oct 02 09:40:48 crc kubenswrapper[5035]: E1002 09:40:48.969307 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="registry-server" Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.969325 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="registry-server" Oct 02 09:40:48 crc kubenswrapper[5035]: E1002 09:40:48.969444 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="extract-utilities" Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.969455 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="extract-utilities" Oct 02 09:40:48 crc kubenswrapper[5035]: E1002 09:40:48.969473 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="extract-content" Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.969480 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="extract-content" Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.971639 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c6b787b-3b9d-49d4-a51f-cd4a98f348e1" containerName="registry-server" Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.973274 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bn74v"] Oct 02 09:40:48 crc kubenswrapper[5035]: I1002 09:40:48.973386 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.102660 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-utilities\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.102759 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-catalog-content\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.102819 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9gl6\" (UniqueName: \"kubernetes.io/projected/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-kube-api-access-q9gl6\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.203461 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-catalog-content\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.203510 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9gl6\" (UniqueName: \"kubernetes.io/projected/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-kube-api-access-q9gl6\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.203600 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-utilities\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.203942 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-catalog-content\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.203987 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-utilities\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.225649 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9gl6\" (UniqueName: \"kubernetes.io/projected/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-kube-api-access-q9gl6\") pod \"certified-operators-bn74v\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:49 crc kubenswrapper[5035]: I1002 09:40:49.316147 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:51 crc kubenswrapper[5035]: I1002 09:40:51.392260 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bn74v"] Oct 02 09:40:51 crc kubenswrapper[5035]: I1002 09:40:51.510108 5035 generic.go:334] "Generic (PLEG): container finished" podID="4536e046-502a-4344-87d4-d05e53f3a73b" containerID="e08cf09cc5f8097cbf554e6f6e751bab56b330f1ae9b349702fe2559962ffda7" exitCode=0 Oct 02 09:40:51 crc kubenswrapper[5035]: I1002 09:40:51.510145 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerDied","Data":"e08cf09cc5f8097cbf554e6f6e751bab56b330f1ae9b349702fe2559962ffda7"} Oct 02 09:40:51 crc kubenswrapper[5035]: I1002 09:40:51.512293 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" event={"ID":"3b2dae08-474c-4f53-802a-35133f5f5119","Type":"ContainerStarted","Data":"eb096aae90f693f118df169fb9569d0e8715f509a3a2ac61fae84ab62114b815"} Oct 02 09:40:51 crc kubenswrapper[5035]: I1002 09:40:51.512407 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:40:51 crc kubenswrapper[5035]: I1002 09:40:51.513781 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn74v" event={"ID":"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279","Type":"ContainerStarted","Data":"a3b155d65b18c04a56550b280dff1641a1b8d0738aa13ca597461b2bedb4b715"} Oct 02 09:40:51 crc kubenswrapper[5035]: I1002 09:40:51.549439 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" podStartSLOduration=1.737960992 podStartE2EDuration="8.549415423s" podCreationTimestamp="2025-10-02 09:40:43 +0000 UTC" firstStartedPulling="2025-10-02 09:40:44.361847233 +0000 UTC m=+809.718191288" lastFinishedPulling="2025-10-02 09:40:51.173301684 +0000 UTC m=+816.529645719" observedRunningTime="2025-10-02 09:40:51.547998572 +0000 UTC m=+816.904342597" watchObservedRunningTime="2025-10-02 09:40:51.549415423 +0000 UTC m=+816.905759448" Oct 02 09:40:52 crc kubenswrapper[5035]: I1002 09:40:52.519824 5035 generic.go:334] "Generic (PLEG): container finished" podID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerID="8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d" exitCode=0 Oct 02 09:40:52 crc kubenswrapper[5035]: I1002 09:40:52.519885 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn74v" event={"ID":"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279","Type":"ContainerDied","Data":"8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d"} Oct 02 09:40:52 crc kubenswrapper[5035]: I1002 09:40:52.522823 5035 generic.go:334] "Generic (PLEG): container finished" podID="4536e046-502a-4344-87d4-d05e53f3a73b" containerID="30206ad4a30b2f5a05748e34cd61bbda90b05dc5c63e888a5e76f53f50fd308b" exitCode=0 Oct 02 09:40:52 crc kubenswrapper[5035]: I1002 09:40:52.522890 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerDied","Data":"30206ad4a30b2f5a05748e34cd61bbda90b05dc5c63e888a5e76f53f50fd308b"} Oct 02 09:40:53 crc kubenswrapper[5035]: I1002 09:40:53.435944 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-sms6d" Oct 02 09:40:53 crc kubenswrapper[5035]: I1002 09:40:53.529642 5035 generic.go:334] "Generic (PLEG): container finished" podID="4536e046-502a-4344-87d4-d05e53f3a73b" containerID="74224422de2d5b9cac40f5dd4bb23cc795a2e222cff8122330dc0b11a6d5f192" exitCode=0 Oct 02 09:40:53 crc kubenswrapper[5035]: I1002 09:40:53.529689 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerDied","Data":"74224422de2d5b9cac40f5dd4bb23cc795a2e222cff8122330dc0b11a6d5f192"} Oct 02 09:40:54 crc kubenswrapper[5035]: I1002 09:40:54.539851 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerStarted","Data":"9228e465cc1b2e9e70000792c1b7808ce12d98759ec210822e405d1f124e3f94"} Oct 02 09:40:54 crc kubenswrapper[5035]: I1002 09:40:54.540171 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerStarted","Data":"1b7cfe355e6b1cbe2d78f1e2bb951d54f1a1f595b1a1d599ee1e47612aef9a35"} Oct 02 09:40:54 crc kubenswrapper[5035]: I1002 09:40:54.540187 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerStarted","Data":"3504440e479a6f06d2345bb7cfda8599bece2a00e1659b47c315537e30e35b3f"} Oct 02 09:40:54 crc kubenswrapper[5035]: I1002 09:40:54.540196 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerStarted","Data":"0f086308141829fdba8a1327937e75f023668ad6b96a438d295721b6e6766d0b"} Oct 02 09:40:54 crc kubenswrapper[5035]: I1002 09:40:54.542032 5035 generic.go:334] "Generic (PLEG): container finished" podID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerID="38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960" exitCode=0 Oct 02 09:40:54 crc kubenswrapper[5035]: I1002 09:40:54.542081 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn74v" event={"ID":"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279","Type":"ContainerDied","Data":"38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960"} Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.538042 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.538371 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.538430 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.539075 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f4d0414702a1d48c4861b01b90ad79bd9fac74730456a0cf8ab31e26b08e9e73"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.539134 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://f4d0414702a1d48c4861b01b90ad79bd9fac74730456a0cf8ab31e26b08e9e73" gracePeriod=600 Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.554154 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerStarted","Data":"da62ac7c70c8b2d760708630d73480ea57c19f064d0bee573842ca29945921c9"} Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.554196 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-576rd" event={"ID":"4536e046-502a-4344-87d4-d05e53f3a73b","Type":"ContainerStarted","Data":"8847db86a7801bc7b200acd2a825a928e29ba2c2c1236436026346a18c531396"} Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.554555 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:55 crc kubenswrapper[5035]: I1002 09:40:55.578773 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-576rd" podStartSLOduration=4.941229353 podStartE2EDuration="12.578745733s" podCreationTimestamp="2025-10-02 09:40:43 +0000 UTC" firstStartedPulling="2025-10-02 09:40:43.506263744 +0000 UTC m=+808.862607769" lastFinishedPulling="2025-10-02 09:40:51.143780124 +0000 UTC m=+816.500124149" observedRunningTime="2025-10-02 09:40:55.577326462 +0000 UTC m=+820.933670497" watchObservedRunningTime="2025-10-02 09:40:55.578745733 +0000 UTC m=+820.935089768" Oct 02 09:40:56 crc kubenswrapper[5035]: I1002 09:40:56.567046 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="f4d0414702a1d48c4861b01b90ad79bd9fac74730456a0cf8ab31e26b08e9e73" exitCode=0 Oct 02 09:40:56 crc kubenswrapper[5035]: I1002 09:40:56.567132 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"f4d0414702a1d48c4861b01b90ad79bd9fac74730456a0cf8ab31e26b08e9e73"} Oct 02 09:40:56 crc kubenswrapper[5035]: I1002 09:40:56.567707 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"a1bca98fb1c6a4355a4320624f5591fb78dba746011bbeaf90858ccf387d2ab3"} Oct 02 09:40:56 crc kubenswrapper[5035]: I1002 09:40:56.567735 5035 scope.go:117] "RemoveContainer" containerID="e2879dc2fe72fc93981ca75e0f488799e1bf9edefc894dd701b9e1403934e161" Oct 02 09:40:56 crc kubenswrapper[5035]: I1002 09:40:56.573200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn74v" event={"ID":"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279","Type":"ContainerStarted","Data":"8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105"} Oct 02 09:40:56 crc kubenswrapper[5035]: I1002 09:40:56.607929 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bn74v" podStartSLOduration=5.476722965 podStartE2EDuration="8.607908447s" podCreationTimestamp="2025-10-02 09:40:48 +0000 UTC" firstStartedPulling="2025-10-02 09:40:52.522240986 +0000 UTC m=+817.878585011" lastFinishedPulling="2025-10-02 09:40:55.653426468 +0000 UTC m=+821.009770493" observedRunningTime="2025-10-02 09:40:56.606058053 +0000 UTC m=+821.962402078" watchObservedRunningTime="2025-10-02 09:40:56.607908447 +0000 UTC m=+821.964252472" Oct 02 09:40:58 crc kubenswrapper[5035]: I1002 09:40:58.352838 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:58 crc kubenswrapper[5035]: I1002 09:40:58.418270 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-576rd" Oct 02 09:40:59 crc kubenswrapper[5035]: I1002 09:40:59.317274 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:59 crc kubenswrapper[5035]: I1002 09:40:59.317351 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:40:59 crc kubenswrapper[5035]: I1002 09:40:59.382098 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:41:03 crc kubenswrapper[5035]: I1002 09:41:03.356517 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-576rd" Oct 02 09:41:03 crc kubenswrapper[5035]: I1002 09:41:03.939691 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9qrxg" Oct 02 09:41:04 crc kubenswrapper[5035]: I1002 09:41:04.926091 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-kn2pw" Oct 02 09:41:07 crc kubenswrapper[5035]: I1002 09:41:07.860924 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-966rr"] Oct 02 09:41:07 crc kubenswrapper[5035]: I1002 09:41:07.862387 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-966rr" Oct 02 09:41:07 crc kubenswrapper[5035]: I1002 09:41:07.865652 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 09:41:07 crc kubenswrapper[5035]: I1002 09:41:07.865690 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 09:41:07 crc kubenswrapper[5035]: I1002 09:41:07.870793 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-966rr"] Oct 02 09:41:07 crc kubenswrapper[5035]: I1002 09:41:07.940408 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmjpj\" (UniqueName: \"kubernetes.io/projected/769668d9-804e-49bc-845c-d313b7f6f8a1-kube-api-access-jmjpj\") pod \"openstack-operator-index-966rr\" (UID: \"769668d9-804e-49bc-845c-d313b7f6f8a1\") " pod="openstack-operators/openstack-operator-index-966rr" Oct 02 09:41:08 crc kubenswrapper[5035]: I1002 09:41:08.041997 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmjpj\" (UniqueName: \"kubernetes.io/projected/769668d9-804e-49bc-845c-d313b7f6f8a1-kube-api-access-jmjpj\") pod \"openstack-operator-index-966rr\" (UID: \"769668d9-804e-49bc-845c-d313b7f6f8a1\") " pod="openstack-operators/openstack-operator-index-966rr" Oct 02 09:41:08 crc kubenswrapper[5035]: I1002 09:41:08.061474 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmjpj\" (UniqueName: \"kubernetes.io/projected/769668d9-804e-49bc-845c-d313b7f6f8a1-kube-api-access-jmjpj\") pod \"openstack-operator-index-966rr\" (UID: \"769668d9-804e-49bc-845c-d313b7f6f8a1\") " pod="openstack-operators/openstack-operator-index-966rr" Oct 02 09:41:08 crc kubenswrapper[5035]: I1002 09:41:08.183921 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-966rr" Oct 02 09:41:08 crc kubenswrapper[5035]: I1002 09:41:08.580905 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-966rr"] Oct 02 09:41:08 crc kubenswrapper[5035]: I1002 09:41:08.666293 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-966rr" event={"ID":"769668d9-804e-49bc-845c-d313b7f6f8a1","Type":"ContainerStarted","Data":"83cf4d4b967ada55024cece5e34778b8f50d0b559c05d37cd05c73edd02b84e4"} Oct 02 09:41:09 crc kubenswrapper[5035]: I1002 09:41:09.357545 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:41:11 crc kubenswrapper[5035]: I1002 09:41:11.816262 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-966rr"] Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.621696 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-sz275"] Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.622835 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.627469 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-z2rpq" Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.637400 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sz275"] Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.698992 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-966rr" event={"ID":"769668d9-804e-49bc-845c-d313b7f6f8a1","Type":"ContainerStarted","Data":"d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a"} Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.700032 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vchnx\" (UniqueName: \"kubernetes.io/projected/8c739198-67cc-4ffe-92ee-4ec30714256f-kube-api-access-vchnx\") pod \"openstack-operator-index-sz275\" (UID: \"8c739198-67cc-4ffe-92ee-4ec30714256f\") " pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.714968 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-966rr" podStartSLOduration=2.639422568 podStartE2EDuration="5.714942629s" podCreationTimestamp="2025-10-02 09:41:07 +0000 UTC" firstStartedPulling="2025-10-02 09:41:08.594326602 +0000 UTC m=+833.950670627" lastFinishedPulling="2025-10-02 09:41:11.669846663 +0000 UTC m=+837.026190688" observedRunningTime="2025-10-02 09:41:12.710653175 +0000 UTC m=+838.066997200" watchObservedRunningTime="2025-10-02 09:41:12.714942629 +0000 UTC m=+838.071286654" Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.800950 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vchnx\" (UniqueName: \"kubernetes.io/projected/8c739198-67cc-4ffe-92ee-4ec30714256f-kube-api-access-vchnx\") pod \"openstack-operator-index-sz275\" (UID: \"8c739198-67cc-4ffe-92ee-4ec30714256f\") " pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.824234 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vchnx\" (UniqueName: \"kubernetes.io/projected/8c739198-67cc-4ffe-92ee-4ec30714256f-kube-api-access-vchnx\") pod \"openstack-operator-index-sz275\" (UID: \"8c739198-67cc-4ffe-92ee-4ec30714256f\") " pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:12 crc kubenswrapper[5035]: I1002 09:41:12.941999 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:13 crc kubenswrapper[5035]: W1002 09:41:13.334630 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c739198_67cc_4ffe_92ee_4ec30714256f.slice/crio-3927fadf65b859154449b6a12ecab60069a4c06637d842512dcee3f3199b55f9 WatchSource:0}: Error finding container 3927fadf65b859154449b6a12ecab60069a4c06637d842512dcee3f3199b55f9: Status 404 returned error can't find the container with id 3927fadf65b859154449b6a12ecab60069a4c06637d842512dcee3f3199b55f9 Oct 02 09:41:13 crc kubenswrapper[5035]: I1002 09:41:13.334738 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-sz275"] Oct 02 09:41:13 crc kubenswrapper[5035]: I1002 09:41:13.705984 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sz275" event={"ID":"8c739198-67cc-4ffe-92ee-4ec30714256f","Type":"ContainerStarted","Data":"8561da1cb2168aef1dc3c366c3ab8c91b9266213cee2ed2b73df7947761d6599"} Oct 02 09:41:13 crc kubenswrapper[5035]: I1002 09:41:13.706293 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-sz275" event={"ID":"8c739198-67cc-4ffe-92ee-4ec30714256f","Type":"ContainerStarted","Data":"3927fadf65b859154449b6a12ecab60069a4c06637d842512dcee3f3199b55f9"} Oct 02 09:41:13 crc kubenswrapper[5035]: I1002 09:41:13.706142 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-966rr" podUID="769668d9-804e-49bc-845c-d313b7f6f8a1" containerName="registry-server" containerID="cri-o://d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a" gracePeriod=2 Oct 02 09:41:13 crc kubenswrapper[5035]: I1002 09:41:13.725192 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-sz275" podStartSLOduration=1.60247486 podStartE2EDuration="1.725178402s" podCreationTimestamp="2025-10-02 09:41:12 +0000 UTC" firstStartedPulling="2025-10-02 09:41:13.338802053 +0000 UTC m=+838.695146078" lastFinishedPulling="2025-10-02 09:41:13.461505595 +0000 UTC m=+838.817849620" observedRunningTime="2025-10-02 09:41:13.723725219 +0000 UTC m=+839.080069244" watchObservedRunningTime="2025-10-02 09:41:13.725178402 +0000 UTC m=+839.081522427" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.018339 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bn74v"] Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.018581 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bn74v" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="registry-server" containerID="cri-o://8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105" gracePeriod=2 Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.157813 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-966rr" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.222569 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmjpj\" (UniqueName: \"kubernetes.io/projected/769668d9-804e-49bc-845c-d313b7f6f8a1-kube-api-access-jmjpj\") pod \"769668d9-804e-49bc-845c-d313b7f6f8a1\" (UID: \"769668d9-804e-49bc-845c-d313b7f6f8a1\") " Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.230837 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/769668d9-804e-49bc-845c-d313b7f6f8a1-kube-api-access-jmjpj" (OuterVolumeSpecName: "kube-api-access-jmjpj") pod "769668d9-804e-49bc-845c-d313b7f6f8a1" (UID: "769668d9-804e-49bc-845c-d313b7f6f8a1"). InnerVolumeSpecName "kube-api-access-jmjpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.324774 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmjpj\" (UniqueName: \"kubernetes.io/projected/769668d9-804e-49bc-845c-d313b7f6f8a1-kube-api-access-jmjpj\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.374023 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.425781 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-utilities\") pod \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.425897 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-catalog-content\") pod \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.425955 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9gl6\" (UniqueName: \"kubernetes.io/projected/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-kube-api-access-q9gl6\") pod \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\" (UID: \"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279\") " Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.426877 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-utilities" (OuterVolumeSpecName: "utilities") pod "c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" (UID: "c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.429906 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-kube-api-access-q9gl6" (OuterVolumeSpecName: "kube-api-access-q9gl6") pod "c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" (UID: "c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279"). InnerVolumeSpecName "kube-api-access-q9gl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.469259 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" (UID: "c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.527279 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9gl6\" (UniqueName: \"kubernetes.io/projected/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-kube-api-access-q9gl6\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.527321 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.527333 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.713392 5035 generic.go:334] "Generic (PLEG): container finished" podID="769668d9-804e-49bc-845c-d313b7f6f8a1" containerID="d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a" exitCode=0 Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.713476 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-966rr" event={"ID":"769668d9-804e-49bc-845c-d313b7f6f8a1","Type":"ContainerDied","Data":"d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a"} Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.713497 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-966rr" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.713554 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-966rr" event={"ID":"769668d9-804e-49bc-845c-d313b7f6f8a1","Type":"ContainerDied","Data":"83cf4d4b967ada55024cece5e34778b8f50d0b559c05d37cd05c73edd02b84e4"} Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.713583 5035 scope.go:117] "RemoveContainer" containerID="d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.716413 5035 generic.go:334] "Generic (PLEG): container finished" podID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerID="8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105" exitCode=0 Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.716808 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn74v" event={"ID":"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279","Type":"ContainerDied","Data":"8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105"} Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.716861 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bn74v" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.716859 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bn74v" event={"ID":"c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279","Type":"ContainerDied","Data":"a3b155d65b18c04a56550b280dff1641a1b8d0738aa13ca597461b2bedb4b715"} Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.729923 5035 scope.go:117] "RemoveContainer" containerID="d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a" Oct 02 09:41:14 crc kubenswrapper[5035]: E1002 09:41:14.730358 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a\": container with ID starting with d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a not found: ID does not exist" containerID="d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.730414 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a"} err="failed to get container status \"d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a\": rpc error: code = NotFound desc = could not find container \"d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a\": container with ID starting with d2ba60748b6288c22cc9b09c855b4daee4b3d7153937764407bb334c6c84607a not found: ID does not exist" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.730442 5035 scope.go:117] "RemoveContainer" containerID="8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.749395 5035 scope.go:117] "RemoveContainer" containerID="38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.754318 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-966rr"] Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.761625 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-966rr"] Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.766612 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bn74v"] Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.770300 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bn74v"] Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.771265 5035 scope.go:117] "RemoveContainer" containerID="8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.783944 5035 scope.go:117] "RemoveContainer" containerID="8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105" Oct 02 09:41:14 crc kubenswrapper[5035]: E1002 09:41:14.784502 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105\": container with ID starting with 8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105 not found: ID does not exist" containerID="8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.784582 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105"} err="failed to get container status \"8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105\": rpc error: code = NotFound desc = could not find container \"8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105\": container with ID starting with 8593addffa3369c576ac7539a61ff31ee274ebf83fc8d8f9f015cdd128066105 not found: ID does not exist" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.784611 5035 scope.go:117] "RemoveContainer" containerID="38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960" Oct 02 09:41:14 crc kubenswrapper[5035]: E1002 09:41:14.784840 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960\": container with ID starting with 38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960 not found: ID does not exist" containerID="38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.784859 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960"} err="failed to get container status \"38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960\": rpc error: code = NotFound desc = could not find container \"38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960\": container with ID starting with 38bfeab362b11423d81a9f40fbacca11b14fde595f4900b8a9e7f590c04c9960 not found: ID does not exist" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.784871 5035 scope.go:117] "RemoveContainer" containerID="8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d" Oct 02 09:41:14 crc kubenswrapper[5035]: E1002 09:41:14.785165 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d\": container with ID starting with 8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d not found: ID does not exist" containerID="8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d" Oct 02 09:41:14 crc kubenswrapper[5035]: I1002 09:41:14.785203 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d"} err="failed to get container status \"8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d\": rpc error: code = NotFound desc = could not find container \"8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d\": container with ID starting with 8acf2a32ceb8dc0bb8e93687b5349ad64efc3a280488c2bc700fac519193213d not found: ID does not exist" Oct 02 09:41:16 crc kubenswrapper[5035]: I1002 09:41:16.171800 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="769668d9-804e-49bc-845c-d313b7f6f8a1" path="/var/lib/kubelet/pods/769668d9-804e-49bc-845c-d313b7f6f8a1/volumes" Oct 02 09:41:16 crc kubenswrapper[5035]: I1002 09:41:16.172550 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" path="/var/lib/kubelet/pods/c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279/volumes" Oct 02 09:41:22 crc kubenswrapper[5035]: I1002 09:41:22.942961 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:22 crc kubenswrapper[5035]: I1002 09:41:22.943565 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:22 crc kubenswrapper[5035]: I1002 09:41:22.972020 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:23 crc kubenswrapper[5035]: I1002 09:41:23.809505 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-sz275" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.257155 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf"] Oct 02 09:41:26 crc kubenswrapper[5035]: E1002 09:41:26.257964 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="registry-server" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.258009 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="registry-server" Oct 02 09:41:26 crc kubenswrapper[5035]: E1002 09:41:26.258024 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="extract-content" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.258033 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="extract-content" Oct 02 09:41:26 crc kubenswrapper[5035]: E1002 09:41:26.258044 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="extract-utilities" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.258051 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="extract-utilities" Oct 02 09:41:26 crc kubenswrapper[5035]: E1002 09:41:26.258063 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769668d9-804e-49bc-845c-d313b7f6f8a1" containerName="registry-server" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.258069 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="769668d9-804e-49bc-845c-d313b7f6f8a1" containerName="registry-server" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.258193 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="769668d9-804e-49bc-845c-d313b7f6f8a1" containerName="registry-server" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.258205 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6b94cc6-c0cc-4e9d-9ff2-e2baebafc279" containerName="registry-server" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.259263 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.261148 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6rnx9" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.267392 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf"] Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.425013 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djw65\" (UniqueName: \"kubernetes.io/projected/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-kube-api-access-djw65\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.425081 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-util\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.425182 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-bundle\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.526879 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djw65\" (UniqueName: \"kubernetes.io/projected/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-kube-api-access-djw65\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.526935 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-util\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.526999 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-bundle\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.527440 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-bundle\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.527596 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-util\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.556893 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djw65\" (UniqueName: \"kubernetes.io/projected/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-kube-api-access-djw65\") pod \"7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:26 crc kubenswrapper[5035]: I1002 09:41:26.633937 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:27 crc kubenswrapper[5035]: I1002 09:41:27.098521 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf"] Oct 02 09:41:27 crc kubenswrapper[5035]: I1002 09:41:27.811625 5035 generic.go:334] "Generic (PLEG): container finished" podID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerID="afdc5cb5e1760afc26fbdf0128bbcde195fadfc2824635b4bbca40c5f440985e" exitCode=0 Oct 02 09:41:27 crc kubenswrapper[5035]: I1002 09:41:27.811967 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" event={"ID":"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8","Type":"ContainerDied","Data":"afdc5cb5e1760afc26fbdf0128bbcde195fadfc2824635b4bbca40c5f440985e"} Oct 02 09:41:27 crc kubenswrapper[5035]: I1002 09:41:27.811996 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" event={"ID":"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8","Type":"ContainerStarted","Data":"5c262a3f909a4c202ce3694c694dd66cd869f8d4b71ce113b8447d9cdea8e420"} Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.431577 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bx4bq"] Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.432806 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.444727 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx4bq"] Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.556160 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-catalog-content\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.556262 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-utilities\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.556289 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ps2z\" (UniqueName: \"kubernetes.io/projected/b431d835-2210-46dc-9206-5cbfbdd8b7c5-kube-api-access-7ps2z\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.657334 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ps2z\" (UniqueName: \"kubernetes.io/projected/b431d835-2210-46dc-9206-5cbfbdd8b7c5-kube-api-access-7ps2z\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.657453 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-catalog-content\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.657524 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-utilities\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.658022 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-utilities\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.658127 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-catalog-content\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.679163 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ps2z\" (UniqueName: \"kubernetes.io/projected/b431d835-2210-46dc-9206-5cbfbdd8b7c5-kube-api-access-7ps2z\") pod \"redhat-marketplace-bx4bq\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.777826 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.825903 5035 generic.go:334] "Generic (PLEG): container finished" podID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerID="636db1f3cfdb5327a3d754a5b80f8feb75bd4e286250b36477cb9161a5d587a6" exitCode=0 Oct 02 09:41:28 crc kubenswrapper[5035]: I1002 09:41:28.825952 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" event={"ID":"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8","Type":"ContainerDied","Data":"636db1f3cfdb5327a3d754a5b80f8feb75bd4e286250b36477cb9161a5d587a6"} Oct 02 09:41:29 crc kubenswrapper[5035]: I1002 09:41:29.221612 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx4bq"] Oct 02 09:41:29 crc kubenswrapper[5035]: I1002 09:41:29.835994 5035 generic.go:334] "Generic (PLEG): container finished" podID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerID="43a0f35076a81ef0beb75278fda527179e2613ed028107641ba5555aca93c4ec" exitCode=0 Oct 02 09:41:29 crc kubenswrapper[5035]: I1002 09:41:29.836091 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" event={"ID":"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8","Type":"ContainerDied","Data":"43a0f35076a81ef0beb75278fda527179e2613ed028107641ba5555aca93c4ec"} Oct 02 09:41:29 crc kubenswrapper[5035]: I1002 09:41:29.838423 5035 generic.go:334] "Generic (PLEG): container finished" podID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerID="51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0" exitCode=0 Oct 02 09:41:29 crc kubenswrapper[5035]: I1002 09:41:29.838473 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx4bq" event={"ID":"b431d835-2210-46dc-9206-5cbfbdd8b7c5","Type":"ContainerDied","Data":"51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0"} Oct 02 09:41:29 crc kubenswrapper[5035]: I1002 09:41:29.838513 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx4bq" event={"ID":"b431d835-2210-46dc-9206-5cbfbdd8b7c5","Type":"ContainerStarted","Data":"ed206f9c87285528f11a14601a0fcfc0b394c69160723bac6e322a775bb5a2bd"} Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.123590 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.288838 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djw65\" (UniqueName: \"kubernetes.io/projected/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-kube-api-access-djw65\") pod \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.288881 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-util\") pod \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.288916 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-bundle\") pod \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\" (UID: \"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8\") " Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.289814 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-bundle" (OuterVolumeSpecName: "bundle") pod "6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" (UID: "6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.293755 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-kube-api-access-djw65" (OuterVolumeSpecName: "kube-api-access-djw65") pod "6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" (UID: "6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8"). InnerVolumeSpecName "kube-api-access-djw65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.313608 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-util" (OuterVolumeSpecName: "util") pod "6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" (UID: "6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.390744 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djw65\" (UniqueName: \"kubernetes.io/projected/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-kube-api-access-djw65\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.390789 5035 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-util\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.390802 5035 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.850206 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" event={"ID":"6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8","Type":"ContainerDied","Data":"5c262a3f909a4c202ce3694c694dd66cd869f8d4b71ce113b8447d9cdea8e420"} Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.850247 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c262a3f909a4c202ce3694c694dd66cd869f8d4b71ce113b8447d9cdea8e420" Oct 02 09:41:31 crc kubenswrapper[5035]: I1002 09:41:31.850266 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf" Oct 02 09:41:32 crc kubenswrapper[5035]: I1002 09:41:32.859441 5035 generic.go:334] "Generic (PLEG): container finished" podID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerID="d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea" exitCode=0 Oct 02 09:41:32 crc kubenswrapper[5035]: I1002 09:41:32.859502 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx4bq" event={"ID":"b431d835-2210-46dc-9206-5cbfbdd8b7c5","Type":"ContainerDied","Data":"d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea"} Oct 02 09:41:34 crc kubenswrapper[5035]: I1002 09:41:34.878662 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx4bq" event={"ID":"b431d835-2210-46dc-9206-5cbfbdd8b7c5","Type":"ContainerStarted","Data":"5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f"} Oct 02 09:41:34 crc kubenswrapper[5035]: I1002 09:41:34.900286 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bx4bq" podStartSLOduration=2.570253068 podStartE2EDuration="6.900264831s" podCreationTimestamp="2025-10-02 09:41:28 +0000 UTC" firstStartedPulling="2025-10-02 09:41:29.841424549 +0000 UTC m=+855.197768574" lastFinishedPulling="2025-10-02 09:41:34.171436312 +0000 UTC m=+859.527780337" observedRunningTime="2025-10-02 09:41:34.894442731 +0000 UTC m=+860.250786766" watchObservedRunningTime="2025-10-02 09:41:34.900264831 +0000 UTC m=+860.256608866" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.392054 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb"] Oct 02 09:41:35 crc kubenswrapper[5035]: E1002 09:41:35.392342 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerName="extract" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.392358 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerName="extract" Oct 02 09:41:35 crc kubenswrapper[5035]: E1002 09:41:35.392379 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerName="util" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.392389 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerName="util" Oct 02 09:41:35 crc kubenswrapper[5035]: E1002 09:41:35.392406 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerName="pull" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.392415 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerName="pull" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.392560 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8" containerName="extract" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.393288 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.407551 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-tc8ph" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.421230 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb"] Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.552982 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lktxj\" (UniqueName: \"kubernetes.io/projected/db3c6e4e-7e62-40dc-91fa-0c79a1eb8010-kube-api-access-lktxj\") pod \"openstack-operator-controller-operator-5578564d9c-6vmnb\" (UID: \"db3c6e4e-7e62-40dc-91fa-0c79a1eb8010\") " pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.654973 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lktxj\" (UniqueName: \"kubernetes.io/projected/db3c6e4e-7e62-40dc-91fa-0c79a1eb8010-kube-api-access-lktxj\") pod \"openstack-operator-controller-operator-5578564d9c-6vmnb\" (UID: \"db3c6e4e-7e62-40dc-91fa-0c79a1eb8010\") " pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.696290 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lktxj\" (UniqueName: \"kubernetes.io/projected/db3c6e4e-7e62-40dc-91fa-0c79a1eb8010-kube-api-access-lktxj\") pod \"openstack-operator-controller-operator-5578564d9c-6vmnb\" (UID: \"db3c6e4e-7e62-40dc-91fa-0c79a1eb8010\") " pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" Oct 02 09:41:35 crc kubenswrapper[5035]: I1002 09:41:35.713127 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" Oct 02 09:41:36 crc kubenswrapper[5035]: I1002 09:41:36.146547 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb"] Oct 02 09:41:36 crc kubenswrapper[5035]: W1002 09:41:36.150132 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb3c6e4e_7e62_40dc_91fa_0c79a1eb8010.slice/crio-b77753c8bc676bdfa55fcc5081fdbd5e597154ea8d268ae63fad3f6aee57b88a WatchSource:0}: Error finding container b77753c8bc676bdfa55fcc5081fdbd5e597154ea8d268ae63fad3f6aee57b88a: Status 404 returned error can't find the container with id b77753c8bc676bdfa55fcc5081fdbd5e597154ea8d268ae63fad3f6aee57b88a Oct 02 09:41:36 crc kubenswrapper[5035]: I1002 09:41:36.894790 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" event={"ID":"db3c6e4e-7e62-40dc-91fa-0c79a1eb8010","Type":"ContainerStarted","Data":"b77753c8bc676bdfa55fcc5081fdbd5e597154ea8d268ae63fad3f6aee57b88a"} Oct 02 09:41:38 crc kubenswrapper[5035]: I1002 09:41:38.778606 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:38 crc kubenswrapper[5035]: I1002 09:41:38.778868 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:38 crc kubenswrapper[5035]: I1002 09:41:38.827690 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:41 crc kubenswrapper[5035]: I1002 09:41:41.926022 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" event={"ID":"db3c6e4e-7e62-40dc-91fa-0c79a1eb8010","Type":"ContainerStarted","Data":"b926c889e5eccb70bcdc4f6c5156595a8cff320330f80dba52b1058f8e3982a5"} Oct 02 09:41:44 crc kubenswrapper[5035]: I1002 09:41:44.953318 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" event={"ID":"db3c6e4e-7e62-40dc-91fa-0c79a1eb8010","Type":"ContainerStarted","Data":"ecfd842b4ec40223d5fde7441848c0f257f2f008f2b202923a3bd00fd8d3d395"} Oct 02 09:41:44 crc kubenswrapper[5035]: I1002 09:41:44.954270 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" Oct 02 09:41:44 crc kubenswrapper[5035]: I1002 09:41:44.989847 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" podStartSLOduration=1.513303782 podStartE2EDuration="9.989830195s" podCreationTimestamp="2025-10-02 09:41:35 +0000 UTC" firstStartedPulling="2025-10-02 09:41:36.152752155 +0000 UTC m=+861.509096180" lastFinishedPulling="2025-10-02 09:41:44.629278578 +0000 UTC m=+869.985622593" observedRunningTime="2025-10-02 09:41:44.984144409 +0000 UTC m=+870.340488444" watchObservedRunningTime="2025-10-02 09:41:44.989830195 +0000 UTC m=+870.346174220" Oct 02 09:41:48 crc kubenswrapper[5035]: I1002 09:41:48.816254 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:48 crc kubenswrapper[5035]: I1002 09:41:48.859191 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx4bq"] Oct 02 09:41:48 crc kubenswrapper[5035]: I1002 09:41:48.988817 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bx4bq" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="registry-server" containerID="cri-o://5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f" gracePeriod=2 Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.455273 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.538977 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-catalog-content\") pod \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.539061 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ps2z\" (UniqueName: \"kubernetes.io/projected/b431d835-2210-46dc-9206-5cbfbdd8b7c5-kube-api-access-7ps2z\") pod \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.539136 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-utilities\") pod \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\" (UID: \"b431d835-2210-46dc-9206-5cbfbdd8b7c5\") " Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.540308 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-utilities" (OuterVolumeSpecName: "utilities") pod "b431d835-2210-46dc-9206-5cbfbdd8b7c5" (UID: "b431d835-2210-46dc-9206-5cbfbdd8b7c5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.544686 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b431d835-2210-46dc-9206-5cbfbdd8b7c5-kube-api-access-7ps2z" (OuterVolumeSpecName: "kube-api-access-7ps2z") pod "b431d835-2210-46dc-9206-5cbfbdd8b7c5" (UID: "b431d835-2210-46dc-9206-5cbfbdd8b7c5"). InnerVolumeSpecName "kube-api-access-7ps2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.552247 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b431d835-2210-46dc-9206-5cbfbdd8b7c5" (UID: "b431d835-2210-46dc-9206-5cbfbdd8b7c5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.641126 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.641163 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b431d835-2210-46dc-9206-5cbfbdd8b7c5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:49 crc kubenswrapper[5035]: I1002 09:41:49.641177 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ps2z\" (UniqueName: \"kubernetes.io/projected/b431d835-2210-46dc-9206-5cbfbdd8b7c5-kube-api-access-7ps2z\") on node \"crc\" DevicePath \"\"" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.001761 5035 generic.go:334] "Generic (PLEG): container finished" podID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerID="5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f" exitCode=0 Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.001854 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx4bq" event={"ID":"b431d835-2210-46dc-9206-5cbfbdd8b7c5","Type":"ContainerDied","Data":"5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f"} Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.001870 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx4bq" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.001915 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx4bq" event={"ID":"b431d835-2210-46dc-9206-5cbfbdd8b7c5","Type":"ContainerDied","Data":"ed206f9c87285528f11a14601a0fcfc0b394c69160723bac6e322a775bb5a2bd"} Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.001962 5035 scope.go:117] "RemoveContainer" containerID="5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.031961 5035 scope.go:117] "RemoveContainer" containerID="d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.037254 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx4bq"] Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.042565 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx4bq"] Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.066005 5035 scope.go:117] "RemoveContainer" containerID="51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.084311 5035 scope.go:117] "RemoveContainer" containerID="5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f" Oct 02 09:41:50 crc kubenswrapper[5035]: E1002 09:41:50.084633 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f\": container with ID starting with 5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f not found: ID does not exist" containerID="5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.084664 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f"} err="failed to get container status \"5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f\": rpc error: code = NotFound desc = could not find container \"5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f\": container with ID starting with 5d02564e84f1b0f1287fe195cf62f7654089663517965f8e6c3e8c7aaac0a54f not found: ID does not exist" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.084683 5035 scope.go:117] "RemoveContainer" containerID="d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea" Oct 02 09:41:50 crc kubenswrapper[5035]: E1002 09:41:50.085215 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea\": container with ID starting with d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea not found: ID does not exist" containerID="d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.085247 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea"} err="failed to get container status \"d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea\": rpc error: code = NotFound desc = could not find container \"d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea\": container with ID starting with d125b0b6afc4ff433d63f724d5abdce56140dc4ebc762ff3b4df5d9ad1018bea not found: ID does not exist" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.085260 5035 scope.go:117] "RemoveContainer" containerID="51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0" Oct 02 09:41:50 crc kubenswrapper[5035]: E1002 09:41:50.085723 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0\": container with ID starting with 51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0 not found: ID does not exist" containerID="51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.085745 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0"} err="failed to get container status \"51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0\": rpc error: code = NotFound desc = could not find container \"51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0\": container with ID starting with 51d7f2f32f07f9a3c41ae3ae971a4d0820256882f933f9685d6d94f3bf8a4ff0 not found: ID does not exist" Oct 02 09:41:50 crc kubenswrapper[5035]: I1002 09:41:50.170701 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" path="/var/lib/kubelet/pods/b431d835-2210-46dc-9206-5cbfbdd8b7c5/volumes" Oct 02 09:41:55 crc kubenswrapper[5035]: I1002 09:41:55.716419 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5578564d9c-6vmnb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.068810 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc"] Oct 02 09:42:12 crc kubenswrapper[5035]: E1002 09:42:12.069709 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.069727 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="extract-utilities" Oct 02 09:42:12 crc kubenswrapper[5035]: E1002 09:42:12.069736 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.069744 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="extract-content" Oct 02 09:42:12 crc kubenswrapper[5035]: E1002 09:42:12.069754 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.069763 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.069901 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b431d835-2210-46dc-9206-5cbfbdd8b7c5" containerName="registry-server" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.070682 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.072800 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-gxzl9" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.077067 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.078045 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.079945 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-mh5xg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.094043 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.099050 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.104083 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.105277 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.109061 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-r2fps" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.117849 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.127766 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.130089 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.133116 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-72pqc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.142285 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9v9zk\" (UniqueName: \"kubernetes.io/projected/bb9ed734-53de-48a2-ac4d-57c513e74c83-kube-api-access-9v9zk\") pod \"barbican-operator-controller-manager-6ff8b75857-qmrqc\" (UID: \"bb9ed734-53de-48a2-ac4d-57c513e74c83\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.142340 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs5gk\" (UniqueName: \"kubernetes.io/projected/107fac49-7084-4524-848e-4ebc95387af7-kube-api-access-xs5gk\") pod \"cinder-operator-controller-manager-644bddb6d8-msx6h\" (UID: \"107fac49-7084-4524-848e-4ebc95387af7\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.174737 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.182658 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.183668 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.187275 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.188497 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.190264 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-dtpqn" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.191503 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4glhc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.192969 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.218278 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.247421 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.269952 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.271903 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.278859 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.284305 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cdvz\" (UniqueName: \"kubernetes.io/projected/219d8519-1035-4127-a7b3-d59730e87ba8-kube-api-access-6cdvz\") pod \"horizon-operator-controller-manager-9f4696d94-5znkc\" (UID: \"219d8519-1035-4127-a7b3-d59730e87ba8\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.285456 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.288052 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-9w65z" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.291616 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.300925 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-n8vvj" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.353537 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.353693 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.358648 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-svnfl" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.317764 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5msd2\" (UniqueName: \"kubernetes.io/projected/7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5-kube-api-access-5msd2\") pod \"heat-operator-controller-manager-5d889d78cf-dtzbf\" (UID: \"7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.358947 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6bdq\" (UniqueName: \"kubernetes.io/projected/938b87e8-ee53-41d1-b79d-bd57940d12cf-kube-api-access-d6bdq\") pod \"glance-operator-controller-manager-84958c4d49-p72mn\" (UID: \"938b87e8-ee53-41d1-b79d-bd57940d12cf\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.359041 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9v9zk\" (UniqueName: \"kubernetes.io/projected/bb9ed734-53de-48a2-ac4d-57c513e74c83-kube-api-access-9v9zk\") pod \"barbican-operator-controller-manager-6ff8b75857-qmrqc\" (UID: \"bb9ed734-53de-48a2-ac4d-57c513e74c83\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.359138 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs5gk\" (UniqueName: \"kubernetes.io/projected/107fac49-7084-4524-848e-4ebc95387af7-kube-api-access-xs5gk\") pod \"cinder-operator-controller-manager-644bddb6d8-msx6h\" (UID: \"107fac49-7084-4524-848e-4ebc95387af7\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.359500 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgphl\" (UniqueName: \"kubernetes.io/projected/04b44aaf-0053-43c2-a2ad-4cb364e6d045-kube-api-access-pgphl\") pod \"designate-operator-controller-manager-84f4f7b77b-sb4zr\" (UID: \"04b44aaf-0053-43c2-a2ad-4cb364e6d045\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.400135 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.402236 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9v9zk\" (UniqueName: \"kubernetes.io/projected/bb9ed734-53de-48a2-ac4d-57c513e74c83-kube-api-access-9v9zk\") pod \"barbican-operator-controller-manager-6ff8b75857-qmrqc\" (UID: \"bb9ed734-53de-48a2-ac4d-57c513e74c83\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.410902 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.425938 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs5gk\" (UniqueName: \"kubernetes.io/projected/107fac49-7084-4524-848e-4ebc95387af7-kube-api-access-xs5gk\") pod \"cinder-operator-controller-manager-644bddb6d8-msx6h\" (UID: \"107fac49-7084-4524-848e-4ebc95387af7\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.451055 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461426 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7frp\" (UniqueName: \"kubernetes.io/projected/3bb0cc56-b9e6-42fe-9122-929e1f7a60e8-kube-api-access-l7frp\") pod \"keystone-operator-controller-manager-5bd55b4bff-hp45m\" (UID: \"3bb0cc56-b9e6-42fe-9122-929e1f7a60e8\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461498 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgphl\" (UniqueName: \"kubernetes.io/projected/04b44aaf-0053-43c2-a2ad-4cb364e6d045-kube-api-access-pgphl\") pod \"designate-operator-controller-manager-84f4f7b77b-sb4zr\" (UID: \"04b44aaf-0053-43c2-a2ad-4cb364e6d045\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461557 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-cert\") pod \"infra-operator-controller-manager-9d6c5db85-hvhzg\" (UID: \"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461584 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cdvz\" (UniqueName: \"kubernetes.io/projected/219d8519-1035-4127-a7b3-d59730e87ba8-kube-api-access-6cdvz\") pod \"horizon-operator-controller-manager-9f4696d94-5znkc\" (UID: \"219d8519-1035-4127-a7b3-d59730e87ba8\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461612 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbrv8\" (UniqueName: \"kubernetes.io/projected/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-kube-api-access-jbrv8\") pod \"infra-operator-controller-manager-9d6c5db85-hvhzg\" (UID: \"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461737 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx8j7\" (UniqueName: \"kubernetes.io/projected/9c08e6d0-af1a-43cc-a898-d345eba038bb-kube-api-access-hx8j7\") pod \"ironic-operator-controller-manager-d87866488-xc6dw\" (UID: \"9c08e6d0-af1a-43cc-a898-d345eba038bb\") " pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5msd2\" (UniqueName: \"kubernetes.io/projected/7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5-kube-api-access-5msd2\") pod \"heat-operator-controller-manager-5d889d78cf-dtzbf\" (UID: \"7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.461809 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6bdq\" (UniqueName: \"kubernetes.io/projected/938b87e8-ee53-41d1-b79d-bd57940d12cf-kube-api-access-d6bdq\") pod \"glance-operator-controller-manager-84958c4d49-p72mn\" (UID: \"938b87e8-ee53-41d1-b79d-bd57940d12cf\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.468787 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.469763 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.499064 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-mc7h2" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.501718 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.522791 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.523769 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.524755 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5msd2\" (UniqueName: \"kubernetes.io/projected/7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5-kube-api-access-5msd2\") pod \"heat-operator-controller-manager-5d889d78cf-dtzbf\" (UID: \"7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.531969 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-2vsd9" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.533397 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgphl\" (UniqueName: \"kubernetes.io/projected/04b44aaf-0053-43c2-a2ad-4cb364e6d045-kube-api-access-pgphl\") pod \"designate-operator-controller-manager-84f4f7b77b-sb4zr\" (UID: \"04b44aaf-0053-43c2-a2ad-4cb364e6d045\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.533616 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cdvz\" (UniqueName: \"kubernetes.io/projected/219d8519-1035-4127-a7b3-d59730e87ba8-kube-api-access-6cdvz\") pod \"horizon-operator-controller-manager-9f4696d94-5znkc\" (UID: \"219d8519-1035-4127-a7b3-d59730e87ba8\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.535278 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.537140 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6bdq\" (UniqueName: \"kubernetes.io/projected/938b87e8-ee53-41d1-b79d-bd57940d12cf-kube-api-access-d6bdq\") pod \"glance-operator-controller-manager-84958c4d49-p72mn\" (UID: \"938b87e8-ee53-41d1-b79d-bd57940d12cf\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.558732 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.561879 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.562747 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7frp\" (UniqueName: \"kubernetes.io/projected/3bb0cc56-b9e6-42fe-9122-929e1f7a60e8-kube-api-access-l7frp\") pod \"keystone-operator-controller-manager-5bd55b4bff-hp45m\" (UID: \"3bb0cc56-b9e6-42fe-9122-929e1f7a60e8\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.562808 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-cert\") pod \"infra-operator-controller-manager-9d6c5db85-hvhzg\" (UID: \"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.562835 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbrv8\" (UniqueName: \"kubernetes.io/projected/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-kube-api-access-jbrv8\") pod \"infra-operator-controller-manager-9d6c5db85-hvhzg\" (UID: \"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.562855 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx8j7\" (UniqueName: \"kubernetes.io/projected/9c08e6d0-af1a-43cc-a898-d345eba038bb-kube-api-access-hx8j7\") pod \"ironic-operator-controller-manager-d87866488-xc6dw\" (UID: \"9c08e6d0-af1a-43cc-a898-d345eba038bb\") " pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" Oct 02 09:42:12 crc kubenswrapper[5035]: E1002 09:42:12.563264 5035 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 02 09:42:12 crc kubenswrapper[5035]: E1002 09:42:12.563309 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-cert podName:2279a6b9-cecc-4b8b-ac52-041a6abe4b9a nodeName:}" failed. No retries permitted until 2025-10-02 09:42:13.063294924 +0000 UTC m=+898.419638949 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-cert") pod "infra-operator-controller-manager-9d6c5db85-hvhzg" (UID: "2279a6b9-cecc-4b8b-ac52-041a6abe4b9a") : secret "infra-operator-webhook-server-cert" not found Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.575702 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.577783 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-n8tsm" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.580808 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.585349 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7frp\" (UniqueName: \"kubernetes.io/projected/3bb0cc56-b9e6-42fe-9122-929e1f7a60e8-kube-api-access-l7frp\") pod \"keystone-operator-controller-manager-5bd55b4bff-hp45m\" (UID: \"3bb0cc56-b9e6-42fe-9122-929e1f7a60e8\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.588766 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbrv8\" (UniqueName: \"kubernetes.io/projected/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-kube-api-access-jbrv8\") pod \"infra-operator-controller-manager-9d6c5db85-hvhzg\" (UID: \"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.592901 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.593914 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.595220 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx8j7\" (UniqueName: \"kubernetes.io/projected/9c08e6d0-af1a-43cc-a898-d345eba038bb-kube-api-access-hx8j7\") pod \"ironic-operator-controller-manager-d87866488-xc6dw\" (UID: \"9c08e6d0-af1a-43cc-a898-d345eba038bb\") " pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.601875 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-69stc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.623713 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.642565 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.644522 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.652780 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-cb7lr" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.663579 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm628\" (UniqueName: \"kubernetes.io/projected/98a2c0aa-2954-4b93-be62-0eec45a1e3c4-kube-api-access-pm628\") pod \"mariadb-operator-controller-manager-88c7-bcnlz\" (UID: \"98a2c0aa-2954-4b93-be62-0eec45a1e3c4\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.663665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfc5r\" (UniqueName: \"kubernetes.io/projected/25bfe9fe-5431-49cc-941d-a2bd795bf7d8-kube-api-access-tfc5r\") pod \"manila-operator-controller-manager-6d68dbc695-6pm9g\" (UID: \"25bfe9fe-5431-49cc-941d-a2bd795bf7d8\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.663727 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v7hz\" (UniqueName: \"kubernetes.io/projected/bf4ed5a3-c0a1-4b98-837b-20de28931ba8-kube-api-access-7v7hz\") pod \"neutron-operator-controller-manager-849d5b9b84-zh2nk\" (UID: \"bf4ed5a3-c0a1-4b98-837b-20de28931ba8\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.676982 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.677056 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.678302 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.686769 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.687629 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.687736 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8kcqx" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.690769 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.702562 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.709363 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.710590 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.710899 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.713311 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-t96kb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.721837 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.723632 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.724711 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.727734 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-qtksg" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.739229 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.748113 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.753414 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.761137 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.762346 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.764484 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfc5r\" (UniqueName: \"kubernetes.io/projected/25bfe9fe-5431-49cc-941d-a2bd795bf7d8-kube-api-access-tfc5r\") pod \"manila-operator-controller-manager-6d68dbc695-6pm9g\" (UID: \"25bfe9fe-5431-49cc-941d-a2bd795bf7d8\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.764711 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v7hz\" (UniqueName: \"kubernetes.io/projected/bf4ed5a3-c0a1-4b98-837b-20de28931ba8-kube-api-access-7v7hz\") pod \"neutron-operator-controller-manager-849d5b9b84-zh2nk\" (UID: \"bf4ed5a3-c0a1-4b98-837b-20de28931ba8\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.764788 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm628\" (UniqueName: \"kubernetes.io/projected/98a2c0aa-2954-4b93-be62-0eec45a1e3c4-kube-api-access-pm628\") pod \"mariadb-operator-controller-manager-88c7-bcnlz\" (UID: \"98a2c0aa-2954-4b93-be62-0eec45a1e3c4\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.764891 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c88ks\" (UniqueName: \"kubernetes.io/projected/1dc0efdb-db90-4467-9255-056281840a4a-kube-api-access-c88ks\") pod \"nova-operator-controller-manager-64cd67b5cb-sgg6c\" (UID: \"1dc0efdb-db90-4467-9255-056281840a4a\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.764959 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kghmn\" (UniqueName: \"kubernetes.io/projected/ac688684-8289-4fe0-a776-14ec42952e8f-kube-api-access-kghmn\") pod \"octavia-operator-controller-manager-7b787867f4-nsrkb\" (UID: \"ac688684-8289-4fe0-a776-14ec42952e8f\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.765565 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.777981 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-44c2w" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.788532 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.788626 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.790989 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jr8vp" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.791899 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.820766 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-msfdb"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.823162 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.823449 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm628\" (UniqueName: \"kubernetes.io/projected/98a2c0aa-2954-4b93-be62-0eec45a1e3c4-kube-api-access-pm628\") pod \"mariadb-operator-controller-manager-88c7-bcnlz\" (UID: \"98a2c0aa-2954-4b93-be62-0eec45a1e3c4\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.858780 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.861635 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v7hz\" (UniqueName: \"kubernetes.io/projected/bf4ed5a3-c0a1-4b98-837b-20de28931ba8-kube-api-access-7v7hz\") pod \"neutron-operator-controller-manager-849d5b9b84-zh2nk\" (UID: \"bf4ed5a3-c0a1-4b98-837b-20de28931ba8\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.865950 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfc5r\" (UniqueName: \"kubernetes.io/projected/25bfe9fe-5431-49cc-941d-a2bd795bf7d8-kube-api-access-tfc5r\") pod \"manila-operator-controller-manager-6d68dbc695-6pm9g\" (UID: \"25bfe9fe-5431-49cc-941d-a2bd795bf7d8\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.867422 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjwdm\" (UniqueName: \"kubernetes.io/projected/b9fa1e3a-3fc9-46e6-a1fb-e51755336945-kube-api-access-rjwdm\") pod \"ovn-operator-controller-manager-9976ff44c-rs2lh\" (UID: \"b9fa1e3a-3fc9-46e6-a1fb-e51755336945\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.867680 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c88ks\" (UniqueName: \"kubernetes.io/projected/1dc0efdb-db90-4467-9255-056281840a4a-kube-api-access-c88ks\") pod \"nova-operator-controller-manager-64cd67b5cb-sgg6c\" (UID: \"1dc0efdb-db90-4467-9255-056281840a4a\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.867799 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kghmn\" (UniqueName: \"kubernetes.io/projected/ac688684-8289-4fe0-a776-14ec42952e8f-kube-api-access-kghmn\") pod \"octavia-operator-controller-manager-7b787867f4-nsrkb\" (UID: \"ac688684-8289-4fe0-a776-14ec42952e8f\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.867909 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlhcq\" (UniqueName: \"kubernetes.io/projected/0b78dcac-905a-43ee-bfaa-ec62206f67ba-kube-api-access-mlhcq\") pod \"placement-operator-controller-manager-589c58c6c-8f4pt\" (UID: \"0b78dcac-905a-43ee-bfaa-ec62206f67ba\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.868051 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8rmk\" (UniqueName: \"kubernetes.io/projected/4f7f3b68-a6ec-42b1-be68-d027bc3b5012-kube-api-access-j8rmk\") pod \"swift-operator-controller-manager-84d6b4b759-2lqs2\" (UID: \"4f7f3b68-a6ec-42b1-be68-d027bc3b5012\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.868196 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db1e3e37-5811-4be9-bbcb-62ae86c895fd-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9grgw\" (UID: \"db1e3e37-5811-4be9-bbcb-62ae86c895fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.868290 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b656c\" (UniqueName: \"kubernetes.io/projected/db1e3e37-5811-4be9-bbcb-62ae86c895fd-kube-api-access-b656c\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9grgw\" (UID: \"db1e3e37-5811-4be9-bbcb-62ae86c895fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.869014 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-xfcxp" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.919795 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.920514 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kghmn\" (UniqueName: \"kubernetes.io/projected/ac688684-8289-4fe0-a776-14ec42952e8f-kube-api-access-kghmn\") pod \"octavia-operator-controller-manager-7b787867f4-nsrkb\" (UID: \"ac688684-8289-4fe0-a776-14ec42952e8f\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.921171 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-msfdb"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.929904 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.934078 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c88ks\" (UniqueName: \"kubernetes.io/projected/1dc0efdb-db90-4467-9255-056281840a4a-kube-api-access-c88ks\") pod \"nova-operator-controller-manager-64cd67b5cb-sgg6c\" (UID: \"1dc0efdb-db90-4467-9255-056281840a4a\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.947593 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.948783 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.951029 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-fc96c" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.954231 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc"] Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.955274 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.970014 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjwdm\" (UniqueName: \"kubernetes.io/projected/b9fa1e3a-3fc9-46e6-a1fb-e51755336945-kube-api-access-rjwdm\") pod \"ovn-operator-controller-manager-9976ff44c-rs2lh\" (UID: \"b9fa1e3a-3fc9-46e6-a1fb-e51755336945\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.970061 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlhcq\" (UniqueName: \"kubernetes.io/projected/0b78dcac-905a-43ee-bfaa-ec62206f67ba-kube-api-access-mlhcq\") pod \"placement-operator-controller-manager-589c58c6c-8f4pt\" (UID: \"0b78dcac-905a-43ee-bfaa-ec62206f67ba\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.970098 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8rmk\" (UniqueName: \"kubernetes.io/projected/4f7f3b68-a6ec-42b1-be68-d027bc3b5012-kube-api-access-j8rmk\") pod \"swift-operator-controller-manager-84d6b4b759-2lqs2\" (UID: \"4f7f3b68-a6ec-42b1-be68-d027bc3b5012\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.970138 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kszkz\" (UniqueName: \"kubernetes.io/projected/1b7f13af-0af5-4ffd-b677-e583068cd598-kube-api-access-kszkz\") pod \"telemetry-operator-controller-manager-b8d54b5d7-49mnw\" (UID: \"1b7f13af-0af5-4ffd-b677-e583068cd598\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.970161 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q44t8\" (UniqueName: \"kubernetes.io/projected/c7a8160d-c732-4736-a1b0-57999db04976-kube-api-access-q44t8\") pod \"test-operator-controller-manager-85777745bb-msfdb\" (UID: \"c7a8160d-c732-4736-a1b0-57999db04976\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.970185 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db1e3e37-5811-4be9-bbcb-62ae86c895fd-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9grgw\" (UID: \"db1e3e37-5811-4be9-bbcb-62ae86c895fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.970211 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b656c\" (UniqueName: \"kubernetes.io/projected/db1e3e37-5811-4be9-bbcb-62ae86c895fd-kube-api-access-b656c\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9grgw\" (UID: \"db1e3e37-5811-4be9-bbcb-62ae86c895fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:12 crc kubenswrapper[5035]: E1002 09:42:12.970877 5035 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 09:42:12 crc kubenswrapper[5035]: E1002 09:42:12.970911 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db1e3e37-5811-4be9-bbcb-62ae86c895fd-cert podName:db1e3e37-5811-4be9-bbcb-62ae86c895fd nodeName:}" failed. No retries permitted until 2025-10-02 09:42:13.470899171 +0000 UTC m=+898.827243196 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db1e3e37-5811-4be9-bbcb-62ae86c895fd-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-9grgw" (UID: "db1e3e37-5811-4be9-bbcb-62ae86c895fd") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.985830 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.992120 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b656c\" (UniqueName: \"kubernetes.io/projected/db1e3e37-5811-4be9-bbcb-62ae86c895fd-kube-api-access-b656c\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9grgw\" (UID: \"db1e3e37-5811-4be9-bbcb-62ae86c895fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.997191 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjwdm\" (UniqueName: \"kubernetes.io/projected/b9fa1e3a-3fc9-46e6-a1fb-e51755336945-kube-api-access-rjwdm\") pod \"ovn-operator-controller-manager-9976ff44c-rs2lh\" (UID: \"b9fa1e3a-3fc9-46e6-a1fb-e51755336945\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" Oct 02 09:42:12 crc kubenswrapper[5035]: I1002 09:42:12.997243 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.001151 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8rmk\" (UniqueName: \"kubernetes.io/projected/4f7f3b68-a6ec-42b1-be68-d027bc3b5012-kube-api-access-j8rmk\") pod \"swift-operator-controller-manager-84d6b4b759-2lqs2\" (UID: \"4f7f3b68-a6ec-42b1-be68-d027bc3b5012\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.001723 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlhcq\" (UniqueName: \"kubernetes.io/projected/0b78dcac-905a-43ee-bfaa-ec62206f67ba-kube-api-access-mlhcq\") pod \"placement-operator-controller-manager-589c58c6c-8f4pt\" (UID: \"0b78dcac-905a-43ee-bfaa-ec62206f67ba\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.010328 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.012454 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-zzcn4" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.012683 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.023830 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.045667 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.046721 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.054289 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-zrlrz" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.054301 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.065198 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.074714 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlccw\" (UniqueName: \"kubernetes.io/projected/099bc72a-55f2-438f-b90c-5e54cc4e7f8e-kube-api-access-mlccw\") pod \"watcher-operator-controller-manager-6b9957f54f-m58vc\" (UID: \"099bc72a-55f2-438f-b90c-5e54cc4e7f8e\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.074789 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kszkz\" (UniqueName: \"kubernetes.io/projected/1b7f13af-0af5-4ffd-b677-e583068cd598-kube-api-access-kszkz\") pod \"telemetry-operator-controller-manager-b8d54b5d7-49mnw\" (UID: \"1b7f13af-0af5-4ffd-b677-e583068cd598\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.074815 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q44t8\" (UniqueName: \"kubernetes.io/projected/c7a8160d-c732-4736-a1b0-57999db04976-kube-api-access-q44t8\") pod \"test-operator-controller-manager-85777745bb-msfdb\" (UID: \"c7a8160d-c732-4736-a1b0-57999db04976\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.074868 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-cert\") pod \"infra-operator-controller-manager-9d6c5db85-hvhzg\" (UID: \"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.077599 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2279a6b9-cecc-4b8b-ac52-041a6abe4b9a-cert\") pod \"infra-operator-controller-manager-9d6c5db85-hvhzg\" (UID: \"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.090102 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.092466 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q44t8\" (UniqueName: \"kubernetes.io/projected/c7a8160d-c732-4736-a1b0-57999db04976-kube-api-access-q44t8\") pod \"test-operator-controller-manager-85777745bb-msfdb\" (UID: \"c7a8160d-c732-4736-a1b0-57999db04976\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.093912 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kszkz\" (UniqueName: \"kubernetes.io/projected/1b7f13af-0af5-4ffd-b677-e583068cd598-kube-api-access-kszkz\") pod \"telemetry-operator-controller-manager-b8d54b5d7-49mnw\" (UID: \"1b7f13af-0af5-4ffd-b677-e583068cd598\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.134913 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.143769 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.156903 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.176241 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a35394d-10b4-408d-8666-780370f58653-cert\") pod \"openstack-operator-controller-manager-65bbb5d7c7-shdfq\" (UID: \"0a35394d-10b4-408d-8666-780370f58653\") " pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.176304 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgfgn\" (UniqueName: \"kubernetes.io/projected/0a35394d-10b4-408d-8666-780370f58653-kube-api-access-pgfgn\") pod \"openstack-operator-controller-manager-65bbb5d7c7-shdfq\" (UID: \"0a35394d-10b4-408d-8666-780370f58653\") " pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.176359 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42dzt\" (UniqueName: \"kubernetes.io/projected/3b85995e-f7f9-42d1-9cd5-c32d2ba5444a-kube-api-access-42dzt\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw\" (UID: \"3b85995e-f7f9-42d1-9cd5-c32d2ba5444a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.176381 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlccw\" (UniqueName: \"kubernetes.io/projected/099bc72a-55f2-438f-b90c-5e54cc4e7f8e-kube-api-access-mlccw\") pod \"watcher-operator-controller-manager-6b9957f54f-m58vc\" (UID: \"099bc72a-55f2-438f-b90c-5e54cc4e7f8e\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.181906 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.197311 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlccw\" (UniqueName: \"kubernetes.io/projected/099bc72a-55f2-438f-b90c-5e54cc4e7f8e-kube-api-access-mlccw\") pod \"watcher-operator-controller-manager-6b9957f54f-m58vc\" (UID: \"099bc72a-55f2-438f-b90c-5e54cc4e7f8e\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.225816 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.265966 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.277329 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgfgn\" (UniqueName: \"kubernetes.io/projected/0a35394d-10b4-408d-8666-780370f58653-kube-api-access-pgfgn\") pod \"openstack-operator-controller-manager-65bbb5d7c7-shdfq\" (UID: \"0a35394d-10b4-408d-8666-780370f58653\") " pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.277463 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42dzt\" (UniqueName: \"kubernetes.io/projected/3b85995e-f7f9-42d1-9cd5-c32d2ba5444a-kube-api-access-42dzt\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw\" (UID: \"3b85995e-f7f9-42d1-9cd5-c32d2ba5444a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.277630 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a35394d-10b4-408d-8666-780370f58653-cert\") pod \"openstack-operator-controller-manager-65bbb5d7c7-shdfq\" (UID: \"0a35394d-10b4-408d-8666-780370f58653\") " pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: E1002 09:42:13.278586 5035 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 09:42:13 crc kubenswrapper[5035]: E1002 09:42:13.278660 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a35394d-10b4-408d-8666-780370f58653-cert podName:0a35394d-10b4-408d-8666-780370f58653 nodeName:}" failed. No retries permitted until 2025-10-02 09:42:13.77862505 +0000 UTC m=+899.134969175 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a35394d-10b4-408d-8666-780370f58653-cert") pod "openstack-operator-controller-manager-65bbb5d7c7-shdfq" (UID: "0a35394d-10b4-408d-8666-780370f58653") : secret "webhook-server-cert" not found Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.282252 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.305338 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.318161 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42dzt\" (UniqueName: \"kubernetes.io/projected/3b85995e-f7f9-42d1-9cd5-c32d2ba5444a-kube-api-access-42dzt\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw\" (UID: \"3b85995e-f7f9-42d1-9cd5-c32d2ba5444a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.319500 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgfgn\" (UniqueName: \"kubernetes.io/projected/0a35394d-10b4-408d-8666-780370f58653-kube-api-access-pgfgn\") pod \"openstack-operator-controller-manager-65bbb5d7c7-shdfq\" (UID: \"0a35394d-10b4-408d-8666-780370f58653\") " pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.445905 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.483269 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db1e3e37-5811-4be9-bbcb-62ae86c895fd-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9grgw\" (UID: \"db1e3e37-5811-4be9-bbcb-62ae86c895fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.489502 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db1e3e37-5811-4be9-bbcb-62ae86c895fd-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9grgw\" (UID: \"db1e3e37-5811-4be9-bbcb-62ae86c895fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.609498 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.617045 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.776217 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw"] Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.791485 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a35394d-10b4-408d-8666-780370f58653-cert\") pod \"openstack-operator-controller-manager-65bbb5d7c7-shdfq\" (UID: \"0a35394d-10b4-408d-8666-780370f58653\") " pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.797598 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a35394d-10b4-408d-8666-780370f58653-cert\") pod \"openstack-operator-controller-manager-65bbb5d7c7-shdfq\" (UID: \"0a35394d-10b4-408d-8666-780370f58653\") " pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.864063 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr"] Oct 02 09:42:13 crc kubenswrapper[5035]: W1002 09:42:13.866929 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04b44aaf_0053_43c2_a2ad_4cb364e6d045.slice/crio-606222b5b45c4cc627664af311b86dad5e9f99d209b10bfbd55ed6666b331cdb WatchSource:0}: Error finding container 606222b5b45c4cc627664af311b86dad5e9f99d209b10bfbd55ed6666b331cdb: Status 404 returned error can't find the container with id 606222b5b45c4cc627664af311b86dad5e9f99d209b10bfbd55ed6666b331cdb Oct 02 09:42:13 crc kubenswrapper[5035]: W1002 09:42:13.868214 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod938b87e8_ee53_41d1_b79d_bd57940d12cf.slice/crio-ec39c6ab707cd5bc1016d1163f7da08cac89ab45045f1b9f1efa6dc32ff27be2 WatchSource:0}: Error finding container ec39c6ab707cd5bc1016d1163f7da08cac89ab45045f1b9f1efa6dc32ff27be2: Status 404 returned error can't find the container with id ec39c6ab707cd5bc1016d1163f7da08cac89ab45045f1b9f1efa6dc32ff27be2 Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.874247 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn"] Oct 02 09:42:13 crc kubenswrapper[5035]: W1002 09:42:13.881968 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bb0cc56_b9e6_42fe_9122_929e1f7a60e8.slice/crio-785fcd6e17aeebc36d4919f1e8fb9a471fa46c9a875066222fbc54793d6b2797 WatchSource:0}: Error finding container 785fcd6e17aeebc36d4919f1e8fb9a471fa46c9a875066222fbc54793d6b2797: Status 404 returned error can't find the container with id 785fcd6e17aeebc36d4919f1e8fb9a471fa46c9a875066222fbc54793d6b2797 Oct 02 09:42:13 crc kubenswrapper[5035]: I1002 09:42:13.883555 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.022023 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.027673 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:14 crc kubenswrapper[5035]: W1002 09:42:14.029566 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98a2c0aa_2954_4b93_be62_0eec45a1e3c4.slice/crio-8ec51fda8a8674fd82fb7f02cc09abcd916c36535d29a526c904917083a7f313 WatchSource:0}: Error finding container 8ec51fda8a8674fd82fb7f02cc09abcd916c36535d29a526c904917083a7f313: Status 404 returned error can't find the container with id 8ec51fda8a8674fd82fb7f02cc09abcd916c36535d29a526c904917083a7f313 Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.030036 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz"] Oct 02 09:42:14 crc kubenswrapper[5035]: W1002 09:42:14.030634 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac688684_8289_4fe0_a776_14ec42952e8f.slice/crio-f83d586f063bcc69ca671dfad51a0864c5a63146eca7fd946067fb1a0be2df6e WatchSource:0}: Error finding container f83d586f063bcc69ca671dfad51a0864c5a63146eca7fd946067fb1a0be2df6e: Status 404 returned error can't find the container with id f83d586f063bcc69ca671dfad51a0864c5a63146eca7fd946067fb1a0be2df6e Oct 02 09:42:14 crc kubenswrapper[5035]: W1002 09:42:14.032426 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ee3db1b_fe71_40b5_8ebb_5cee3301dcd5.slice/crio-dd29db17cf6d8a18704fe373ea2c9c24b48dddc79816d074d595d34b867ce01b WatchSource:0}: Error finding container dd29db17cf6d8a18704fe373ea2c9c24b48dddc79816d074d595d34b867ce01b: Status 404 returned error can't find the container with id dd29db17cf6d8a18704fe373ea2c9c24b48dddc79816d074d595d34b867ce01b Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.035880 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.040719 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh"] Oct 02 09:42:14 crc kubenswrapper[5035]: W1002 09:42:14.044876 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9fa1e3a_3fc9_46e6_a1fb_e51755336945.slice/crio-cdee86bd7fc3c496cec455e064eaca249265941c4aae023655255eb5c859b33d WatchSource:0}: Error finding container cdee86bd7fc3c496cec455e064eaca249265941c4aae023655255eb5c859b33d: Status 404 returned error can't find the container with id cdee86bd7fc3c496cec455e064eaca249265941c4aae023655255eb5c859b33d Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.246080 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" event={"ID":"98a2c0aa-2954-4b93-be62-0eec45a1e3c4","Type":"ContainerStarted","Data":"8ec51fda8a8674fd82fb7f02cc09abcd916c36535d29a526c904917083a7f313"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.253268 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" event={"ID":"bb9ed734-53de-48a2-ac4d-57c513e74c83","Type":"ContainerStarted","Data":"ac2d7adf78642d9c79e965d3ab864faad37f3842dd81bb5b48d8a98f72d4c034"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.256796 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" event={"ID":"b9fa1e3a-3fc9-46e6-a1fb-e51755336945","Type":"ContainerStarted","Data":"cdee86bd7fc3c496cec455e064eaca249265941c4aae023655255eb5c859b33d"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.257802 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" event={"ID":"3bb0cc56-b9e6-42fe-9122-929e1f7a60e8","Type":"ContainerStarted","Data":"785fcd6e17aeebc36d4919f1e8fb9a471fa46c9a875066222fbc54793d6b2797"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.258719 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" event={"ID":"9c08e6d0-af1a-43cc-a898-d345eba038bb","Type":"ContainerStarted","Data":"b92481a5377850ebce7ca566afa6661163b1109968edd66ba814c9d6265d7744"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.259815 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" event={"ID":"04b44aaf-0053-43c2-a2ad-4cb364e6d045","Type":"ContainerStarted","Data":"606222b5b45c4cc627664af311b86dad5e9f99d209b10bfbd55ed6666b331cdb"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.260618 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" event={"ID":"107fac49-7084-4524-848e-4ebc95387af7","Type":"ContainerStarted","Data":"709ebf26ac9d5aefaadaec8a52bb017e88187eba5cc6b85b6a0f62b78964ccd0"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.262175 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" event={"ID":"ac688684-8289-4fe0-a776-14ec42952e8f","Type":"ContainerStarted","Data":"f83d586f063bcc69ca671dfad51a0864c5a63146eca7fd946067fb1a0be2df6e"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.262869 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" event={"ID":"219d8519-1035-4127-a7b3-d59730e87ba8","Type":"ContainerStarted","Data":"9a6c5354889b8a0e46c3891e7af01201f485f168694955eee67193b08ac0b6e5"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.263859 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" event={"ID":"7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5","Type":"ContainerStarted","Data":"dd29db17cf6d8a18704fe373ea2c9c24b48dddc79816d074d595d34b867ce01b"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.275787 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" event={"ID":"938b87e8-ee53-41d1-b79d-bd57940d12cf","Type":"ContainerStarted","Data":"ec39c6ab707cd5bc1016d1163f7da08cac89ab45045f1b9f1efa6dc32ff27be2"} Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.432804 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.465617 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.491131 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.514776 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.540587 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-msfdb"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.566350 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.609043 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.618680 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.620721 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc"] Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.670299 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw"] Oct 02 09:42:14 crc kubenswrapper[5035]: W1002 09:42:14.688936 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7a8160d_c732_4736_a1b0_57999db04976.slice/crio-4b36c6befb7bb528d41331306638aaa53ccc82f5cabd8af8eed7ea9394ee14a7 WatchSource:0}: Error finding container 4b36c6befb7bb528d41331306638aaa53ccc82f5cabd8af8eed7ea9394ee14a7: Status 404 returned error can't find the container with id 4b36c6befb7bb528d41331306638aaa53ccc82f5cabd8af8eed7ea9394ee14a7 Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.702437 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw"] Oct 02 09:42:14 crc kubenswrapper[5035]: W1002 09:42:14.716698 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b85995e_f7f9_42d1_9cd5_c32d2ba5444a.slice/crio-7fc2f7f5d6375fe4e970323b7d8c42bf6aeeca7ed45fb9fa04825c3ffabba419 WatchSource:0}: Error finding container 7fc2f7f5d6375fe4e970323b7d8c42bf6aeeca7ed45fb9fa04825c3ffabba419: Status 404 returned error can't find the container with id 7fc2f7f5d6375fe4e970323b7d8c42bf6aeeca7ed45fb9fa04825c3ffabba419 Oct 02 09:42:14 crc kubenswrapper[5035]: I1002 09:42:14.730435 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq"] Oct 02 09:42:14 crc kubenswrapper[5035]: E1002 09:42:14.745312 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c88ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-sgg6c_openstack-operators(1dc0efdb-db90-4467-9255-056281840a4a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:42:14 crc kubenswrapper[5035]: E1002 09:42:14.745412 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tfc5r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6d68dbc695-6pm9g_openstack-operators(25bfe9fe-5431-49cc-941d-a2bd795bf7d8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:42:14 crc kubenswrapper[5035]: E1002 09:42:14.745499 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mlhcq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-8f4pt_openstack-operators(0b78dcac-905a-43ee-bfaa-ec62206f67ba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:42:14 crc kubenswrapper[5035]: E1002 09:42:14.745512 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-42dzt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw_openstack-operators(3b85995e-f7f9-42d1-9cd5-c32d2ba5444a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:42:14 crc kubenswrapper[5035]: E1002 09:42:14.746873 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" podUID="3b85995e-f7f9-42d1-9cd5-c32d2ba5444a" Oct 02 09:42:14 crc kubenswrapper[5035]: E1002 09:42:14.751812 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b656c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5869cb545-9grgw_openstack-operators(db1e3e37-5811-4be9-bbcb-62ae86c895fd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.283740 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" event={"ID":"1b7f13af-0af5-4ffd-b677-e583068cd598","Type":"ContainerStarted","Data":"298deef266aa22aeaaab04cf545d3cc313d8640b043d55e6876bd3c781cd9c66"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.285623 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" event={"ID":"bf4ed5a3-c0a1-4b98-837b-20de28931ba8","Type":"ContainerStarted","Data":"79330ef63a5289afb8f00fb2a205b37422971222a88ac64ea40ed9f1a8770eb8"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.286656 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" event={"ID":"0a35394d-10b4-408d-8666-780370f58653","Type":"ContainerStarted","Data":"cb4f8f6335ebf2cef943a9b83d5ba18df15a2841e532323cfcbee66041ca91ed"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.287830 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" event={"ID":"db1e3e37-5811-4be9-bbcb-62ae86c895fd","Type":"ContainerStarted","Data":"9bce398d9a49c1b04b62e0f8709de5b3f3708e9bbf94a777792b99e05b6bbdf0"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.289010 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" event={"ID":"c7a8160d-c732-4736-a1b0-57999db04976","Type":"ContainerStarted","Data":"4b36c6befb7bb528d41331306638aaa53ccc82f5cabd8af8eed7ea9394ee14a7"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.290011 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" event={"ID":"099bc72a-55f2-438f-b90c-5e54cc4e7f8e","Type":"ContainerStarted","Data":"3b5b07f94abfc12671a5536f0e42b956a980fc1337e41cda62251bef255dc52d"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.297170 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" event={"ID":"1dc0efdb-db90-4467-9255-056281840a4a","Type":"ContainerStarted","Data":"29d9be1f9a1f82ca32f22d1cfdb6d65da67cc54c0d0c7dea28b99b1f71dba8ed"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.298833 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" event={"ID":"3b85995e-f7f9-42d1-9cd5-c32d2ba5444a","Type":"ContainerStarted","Data":"7fc2f7f5d6375fe4e970323b7d8c42bf6aeeca7ed45fb9fa04825c3ffabba419"} Oct 02 09:42:15 crc kubenswrapper[5035]: E1002 09:42:15.300337 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" podUID="3b85995e-f7f9-42d1-9cd5-c32d2ba5444a" Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.302494 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" event={"ID":"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a","Type":"ContainerStarted","Data":"b77b07b617253c6d34d89ff7238943a558ad881b33f505fb4ea5131386713d90"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.304607 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" event={"ID":"4f7f3b68-a6ec-42b1-be68-d027bc3b5012","Type":"ContainerStarted","Data":"36938829d4be3d5385ddd1aae6a2919f06735f60fd98eb487bca9741ff46a7a6"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.307827 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" event={"ID":"25bfe9fe-5431-49cc-941d-a2bd795bf7d8","Type":"ContainerStarted","Data":"715d4be785daf76ad53fa51b67afd8c862b037553291e63dc27dbcdd06e396f3"} Oct 02 09:42:15 crc kubenswrapper[5035]: I1002 09:42:15.312471 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" event={"ID":"0b78dcac-905a-43ee-bfaa-ec62206f67ba","Type":"ContainerStarted","Data":"d2f0e6931f1d635d9f759a32a0c38b4f7a575bf0ee9d36bcf1d7934e250d179d"} Oct 02 09:42:15 crc kubenswrapper[5035]: E1002 09:42:15.713326 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" podUID="db1e3e37-5811-4be9-bbcb-62ae86c895fd" Oct 02 09:42:15 crc kubenswrapper[5035]: E1002 09:42:15.715766 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" podUID="0b78dcac-905a-43ee-bfaa-ec62206f67ba" Oct 02 09:42:15 crc kubenswrapper[5035]: E1002 09:42:15.715952 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" podUID="25bfe9fe-5431-49cc-941d-a2bd795bf7d8" Oct 02 09:42:15 crc kubenswrapper[5035]: E1002 09:42:15.716039 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" podUID="1dc0efdb-db90-4467-9255-056281840a4a" Oct 02 09:42:16 crc kubenswrapper[5035]: I1002 09:42:16.318876 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" event={"ID":"25bfe9fe-5431-49cc-941d-a2bd795bf7d8","Type":"ContainerStarted","Data":"e398838f9d1ffb234c681d470b5a36248b07ebc2411e01404fb98960d466481f"} Oct 02 09:42:16 crc kubenswrapper[5035]: E1002 09:42:16.320491 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" podUID="25bfe9fe-5431-49cc-941d-a2bd795bf7d8" Oct 02 09:42:16 crc kubenswrapper[5035]: E1002 09:42:16.321637 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" podUID="0b78dcac-905a-43ee-bfaa-ec62206f67ba" Oct 02 09:42:16 crc kubenswrapper[5035]: I1002 09:42:16.320872 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" event={"ID":"0b78dcac-905a-43ee-bfaa-ec62206f67ba","Type":"ContainerStarted","Data":"d3e54a026129a179e040360240a8a13d515d2eb1fad48e82cf62e9b33b4030bc"} Oct 02 09:42:16 crc kubenswrapper[5035]: I1002 09:42:16.322462 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" event={"ID":"db1e3e37-5811-4be9-bbcb-62ae86c895fd","Type":"ContainerStarted","Data":"54c644ad06dc79fecf6f27d9c817c0f50c4d302f3bee155191ffb8a13ada5854"} Oct 02 09:42:16 crc kubenswrapper[5035]: E1002 09:42:16.323692 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" podUID="db1e3e37-5811-4be9-bbcb-62ae86c895fd" Oct 02 09:42:16 crc kubenswrapper[5035]: I1002 09:42:16.323821 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" event={"ID":"1dc0efdb-db90-4467-9255-056281840a4a","Type":"ContainerStarted","Data":"eb1112447223b11f1c703b61380f0fb9dce9a3f155dfa18c358c8e18cd2ce750"} Oct 02 09:42:16 crc kubenswrapper[5035]: E1002 09:42:16.324682 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" podUID="1dc0efdb-db90-4467-9255-056281840a4a" Oct 02 09:42:16 crc kubenswrapper[5035]: I1002 09:42:16.329394 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" event={"ID":"0a35394d-10b4-408d-8666-780370f58653","Type":"ContainerStarted","Data":"e17ccbf45371e8140a5e5570928d2953b1bf7defcb330439a07c91ad46b34ec0"} Oct 02 09:42:16 crc kubenswrapper[5035]: I1002 09:42:16.329470 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" event={"ID":"0a35394d-10b4-408d-8666-780370f58653","Type":"ContainerStarted","Data":"aa4283a1c9497367853cf829c681db86bd5b0d40276a92e363b417ed516a449f"} Oct 02 09:42:16 crc kubenswrapper[5035]: E1002 09:42:16.330978 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" podUID="3b85995e-f7f9-42d1-9cd5-c32d2ba5444a" Oct 02 09:42:16 crc kubenswrapper[5035]: I1002 09:42:16.808638 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" podStartSLOduration=4.808611151 podStartE2EDuration="4.808611151s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:42:16.803922385 +0000 UTC m=+902.160266410" watchObservedRunningTime="2025-10-02 09:42:16.808611151 +0000 UTC m=+902.164955176" Oct 02 09:42:17 crc kubenswrapper[5035]: I1002 09:42:17.335777 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:17 crc kubenswrapper[5035]: E1002 09:42:17.337369 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" podUID="0b78dcac-905a-43ee-bfaa-ec62206f67ba" Oct 02 09:42:17 crc kubenswrapper[5035]: E1002 09:42:17.337745 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" podUID="1dc0efdb-db90-4467-9255-056281840a4a" Oct 02 09:42:17 crc kubenswrapper[5035]: E1002 09:42:17.337798 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" podUID="db1e3e37-5811-4be9-bbcb-62ae86c895fd" Oct 02 09:42:17 crc kubenswrapper[5035]: E1002 09:42:17.337558 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" podUID="25bfe9fe-5431-49cc-941d-a2bd795bf7d8" Oct 02 09:42:20 crc kubenswrapper[5035]: I1002 09:42:20.883054 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zqrs8"] Oct 02 09:42:20 crc kubenswrapper[5035]: I1002 09:42:20.885317 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:20 crc kubenswrapper[5035]: I1002 09:42:20.907583 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqrs8"] Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.036900 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-utilities\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.037218 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-catalog-content\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.037353 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcrl2\" (UniqueName: \"kubernetes.io/projected/8b418acc-8124-4b14-bb79-c036405f371b-kube-api-access-rcrl2\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.138387 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-utilities\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.138448 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-catalog-content\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.138486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcrl2\" (UniqueName: \"kubernetes.io/projected/8b418acc-8124-4b14-bb79-c036405f371b-kube-api-access-rcrl2\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.139045 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-utilities\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.139116 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-catalog-content\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.176554 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcrl2\" (UniqueName: \"kubernetes.io/projected/8b418acc-8124-4b14-bb79-c036405f371b-kube-api-access-rcrl2\") pod \"community-operators-zqrs8\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:21 crc kubenswrapper[5035]: I1002 09:42:21.229437 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:24 crc kubenswrapper[5035]: I1002 09:42:24.042689 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-65bbb5d7c7-shdfq" Oct 02 09:42:27 crc kubenswrapper[5035]: E1002 09:42:27.581245 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.214:5001/openstack-k8s-operators/ironic-operator:71da00fcc5e3c433611a22842dfc124c76ad27f7" Oct 02 09:42:27 crc kubenswrapper[5035]: E1002 09:42:27.581828 5035 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.214:5001/openstack-k8s-operators/ironic-operator:71da00fcc5e3c433611a22842dfc124c76ad27f7" Oct 02 09:42:27 crc kubenswrapper[5035]: E1002 09:42:27.581993 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.214:5001/openstack-k8s-operators/ironic-operator:71da00fcc5e3c433611a22842dfc124c76ad27f7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hx8j7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-d87866488-xc6dw_openstack-operators(9c08e6d0-af1a-43cc-a898-d345eba038bb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:42:34 crc kubenswrapper[5035]: I1002 09:42:34.885269 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zqrs8"] Oct 02 09:42:36 crc kubenswrapper[5035]: W1002 09:42:36.512579 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b418acc_8124_4b14_bb79_c036405f371b.slice/crio-08e0f01aced94efaf99196b3030eec1ce2d99a43d4bd2864dd4d39234575295f WatchSource:0}: Error finding container 08e0f01aced94efaf99196b3030eec1ce2d99a43d4bd2864dd4d39234575295f: Status 404 returned error can't find the container with id 08e0f01aced94efaf99196b3030eec1ce2d99a43d4bd2864dd4d39234575295f Oct 02 09:42:36 crc kubenswrapper[5035]: E1002 09:42:36.735310 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" podUID="9c08e6d0-af1a-43cc-a898-d345eba038bb" Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.497011 5035 generic.go:334] "Generic (PLEG): container finished" podID="8b418acc-8124-4b14-bb79-c036405f371b" containerID="71e156ce725f28679f9d21215941390f217f47fa3858c8c275c62dfade25315e" exitCode=0 Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.497833 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqrs8" event={"ID":"8b418acc-8124-4b14-bb79-c036405f371b","Type":"ContainerDied","Data":"71e156ce725f28679f9d21215941390f217f47fa3858c8c275c62dfade25315e"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.497870 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqrs8" event={"ID":"8b418acc-8124-4b14-bb79-c036405f371b","Type":"ContainerStarted","Data":"08e0f01aced94efaf99196b3030eec1ce2d99a43d4bd2864dd4d39234575295f"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.505823 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" event={"ID":"938b87e8-ee53-41d1-b79d-bd57940d12cf","Type":"ContainerStarted","Data":"c7556efcc8fcb0279b8586016ab22ced79bb17defa70002d5bf0638f0ff95fd9"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.507789 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" event={"ID":"04b44aaf-0053-43c2-a2ad-4cb364e6d045","Type":"ContainerStarted","Data":"97c33ed5198d89fabaec616e05789e689a363f5520df41c939b37665d343841f"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.523619 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" event={"ID":"107fac49-7084-4524-848e-4ebc95387af7","Type":"ContainerStarted","Data":"8076a22e546c5e5d87d49ff60db8e1e9065c42905d9d9016eb0a6e4518b04749"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.534728 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" event={"ID":"98a2c0aa-2954-4b93-be62-0eec45a1e3c4","Type":"ContainerStarted","Data":"b150b14413ecb1ff00835552943169908952d5b820e556c01a2bd99fd79114d0"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.552991 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" event={"ID":"3bb0cc56-b9e6-42fe-9122-929e1f7a60e8","Type":"ContainerStarted","Data":"3cedefe82bd6b253a12fb92fc3e7a7b98e83b913c766f0dc749f50cfc38df580"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.564391 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" event={"ID":"4f7f3b68-a6ec-42b1-be68-d027bc3b5012","Type":"ContainerStarted","Data":"0ccab23593e358ba6f8c90d827d3b460c3aa063965f0dc673a59028edbb709ed"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.577748 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" event={"ID":"1b7f13af-0af5-4ffd-b677-e583068cd598","Type":"ContainerStarted","Data":"a209946e46f79157bb49d781c6c6b307315d55e4e57ab21316186e01a131e8c6"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.581981 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" event={"ID":"ac688684-8289-4fe0-a776-14ec42952e8f","Type":"ContainerStarted","Data":"4f04356457e218306a900cd63ecf16b784546d5939d5af02e3a73a8f9a6f8b38"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.583398 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" event={"ID":"9c08e6d0-af1a-43cc-a898-d345eba038bb","Type":"ContainerStarted","Data":"d105095cb7dce256dc9d399267f38251c3123556af3a96df933c1ba357828cc3"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.586520 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" event={"ID":"25bfe9fe-5431-49cc-941d-a2bd795bf7d8","Type":"ContainerStarted","Data":"5f811f27eb799cc36d5dcf94b2ecd1331c48bd515f71615c90d8bfa4fb6208c1"} Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.586708 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" Oct 02 09:42:37 crc kubenswrapper[5035]: I1002 09:42:37.632747 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" podStartSLOduration=3.695763112 podStartE2EDuration="25.632719128s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.745363002 +0000 UTC m=+900.101707027" lastFinishedPulling="2025-10-02 09:42:36.682319018 +0000 UTC m=+922.038663043" observedRunningTime="2025-10-02 09:42:37.625008603 +0000 UTC m=+922.981352628" watchObservedRunningTime="2025-10-02 09:42:37.632719128 +0000 UTC m=+922.989063163" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.614058 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" event={"ID":"099bc72a-55f2-438f-b90c-5e54cc4e7f8e","Type":"ContainerStarted","Data":"27d355fb3db0ead9aa5933bfd0f243009fba6357cd38b15e822de46ca9a6aa3d"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.619857 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" event={"ID":"1dc0efdb-db90-4467-9255-056281840a4a","Type":"ContainerStarted","Data":"9977ce944d51ea3dc64c84c14682f2f1eb9435b93a63fa04918e63beede440e3"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.621230 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.628951 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" event={"ID":"219d8519-1035-4127-a7b3-d59730e87ba8","Type":"ContainerStarted","Data":"84c0b12b18996ea5b9cb3af29209a4d585e423b66319eb2708f868f39fab3e32"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.652054 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" podStartSLOduration=4.717777066 podStartE2EDuration="26.652031133s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.745185717 +0000 UTC m=+900.101529742" lastFinishedPulling="2025-10-02 09:42:36.679439784 +0000 UTC m=+922.035783809" observedRunningTime="2025-10-02 09:42:38.644014309 +0000 UTC m=+924.000358334" watchObservedRunningTime="2025-10-02 09:42:38.652031133 +0000 UTC m=+924.008375168" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.657074 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" event={"ID":"3b85995e-f7f9-42d1-9cd5-c32d2ba5444a","Type":"ContainerStarted","Data":"a50832c070c613fd61ad997ac1a48e1f75f31636c41191e8fdffbeb518e34b73"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.667297 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" event={"ID":"bb9ed734-53de-48a2-ac4d-57c513e74c83","Type":"ContainerStarted","Data":"ea2964e11ea599edb36f106e650a7d394cd293eed6f89c922a8270ae3b49841a"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.670137 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" event={"ID":"b9fa1e3a-3fc9-46e6-a1fb-e51755336945","Type":"ContainerStarted","Data":"86eb4202bdf9d273a0430f09427f76d22f2a19f0cd5132bcac5ecb1bad8a8603"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.672176 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" event={"ID":"0b78dcac-905a-43ee-bfaa-ec62206f67ba","Type":"ContainerStarted","Data":"03d8e7cae694772b2128ac4464312366ff2c2c1f60f29fc81125dff618ae3a43"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.672395 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.673858 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" event={"ID":"7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5","Type":"ContainerStarted","Data":"58e6d120a969289291011face209c54fa2d6cede7f5c55d602b04c717c6f7150"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.680565 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" event={"ID":"db1e3e37-5811-4be9-bbcb-62ae86c895fd","Type":"ContainerStarted","Data":"83551ee3d4b7b2a981ced01b0c0deb99e6eab2c5db3b0be2b06a2190ed29ac5a"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.680946 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.687664 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" event={"ID":"c7a8160d-c732-4736-a1b0-57999db04976","Type":"ContainerStarted","Data":"1c7821c6d3d3e6a9a4a031575d0a5df51b1d8f49af1ac07ffcb5849b047ce57e"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.689484 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw" podStartSLOduration=4.756766321 podStartE2EDuration="26.689469873s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.745461005 +0000 UTC m=+900.101805030" lastFinishedPulling="2025-10-02 09:42:36.678164557 +0000 UTC m=+922.034508582" observedRunningTime="2025-10-02 09:42:38.685412105 +0000 UTC m=+924.041756130" watchObservedRunningTime="2025-10-02 09:42:38.689469873 +0000 UTC m=+924.045813898" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.690252 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" event={"ID":"ac688684-8289-4fe0-a776-14ec42952e8f","Type":"ContainerStarted","Data":"3feae116d36cf0358aea2a1186eb865d2f60861598f1cfdae7885dfc4578f201"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.690402 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.692415 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" event={"ID":"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a","Type":"ContainerStarted","Data":"884d599fc289172940170433f5a3f986bb033a4d160552616475c2f56fb5b648"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.705179 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" event={"ID":"bf4ed5a3-c0a1-4b98-837b-20de28931ba8","Type":"ContainerStarted","Data":"f913601a6193701061150803fc366a8589a8cee0f28ca9c129ecbb5856b3187d"} Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.742326 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" podStartSLOduration=4.816428969 podStartE2EDuration="26.742296921s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.751390018 +0000 UTC m=+900.107734043" lastFinishedPulling="2025-10-02 09:42:36.67725797 +0000 UTC m=+922.033601995" observedRunningTime="2025-10-02 09:42:38.735995747 +0000 UTC m=+924.092339772" watchObservedRunningTime="2025-10-02 09:42:38.742296921 +0000 UTC m=+924.098640946" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.773335 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" podStartSLOduration=4.838069838 podStartE2EDuration="26.773315424s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.745386043 +0000 UTC m=+900.101730058" lastFinishedPulling="2025-10-02 09:42:36.680631619 +0000 UTC m=+922.036975644" observedRunningTime="2025-10-02 09:42:38.766817335 +0000 UTC m=+924.123161360" watchObservedRunningTime="2025-10-02 09:42:38.773315424 +0000 UTC m=+924.129659449" Oct 02 09:42:38 crc kubenswrapper[5035]: I1002 09:42:38.790550 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" podStartSLOduration=4.228371247 podStartE2EDuration="26.790514465s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.036853234 +0000 UTC m=+899.393197249" lastFinishedPulling="2025-10-02 09:42:36.598996452 +0000 UTC m=+921.955340467" observedRunningTime="2025-10-02 09:42:38.785635483 +0000 UTC m=+924.141979518" watchObservedRunningTime="2025-10-02 09:42:38.790514465 +0000 UTC m=+924.146858480" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.713643 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" event={"ID":"938b87e8-ee53-41d1-b79d-bd57940d12cf","Type":"ContainerStarted","Data":"931ba160f6ca697069818064efc6cc0f3d728154ba985ddddefe450eddf7b500"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.715361 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.717182 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" event={"ID":"04b44aaf-0053-43c2-a2ad-4cb364e6d045","Type":"ContainerStarted","Data":"868f0b90b1d7b5df067ec8df2642e077645977cea8a72da1c63d073fc90cef49"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.717491 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.720035 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" event={"ID":"98a2c0aa-2954-4b93-be62-0eec45a1e3c4","Type":"ContainerStarted","Data":"0e783c6607add3127861893838ae1175e3d6e00d9acce64396e935181cbdff83"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.720076 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.721754 5035 generic.go:334] "Generic (PLEG): container finished" podID="8b418acc-8124-4b14-bb79-c036405f371b" containerID="9d6a12fe60aee60d90970d6f950e819744b7e5b86d06affbb815a2cd7e6930e7" exitCode=0 Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.721796 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqrs8" event={"ID":"8b418acc-8124-4b14-bb79-c036405f371b","Type":"ContainerDied","Data":"9d6a12fe60aee60d90970d6f950e819744b7e5b86d06affbb815a2cd7e6930e7"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.724308 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" event={"ID":"b9fa1e3a-3fc9-46e6-a1fb-e51755336945","Type":"ContainerStarted","Data":"7e80bfa54ea4adcecdfe28b2dc7375893c3a4058e028a2a32a6808dda89b9d6f"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.724444 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.729680 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" event={"ID":"4f7f3b68-a6ec-42b1-be68-d027bc3b5012","Type":"ContainerStarted","Data":"852e426630f5498dc885688b05cb806e053b96d4155e6d659ca2648e49162f94"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.730238 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.733646 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" event={"ID":"7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5","Type":"ContainerStarted","Data":"aa6ea86245b36ebe5ace2c1f2ea7c237b94f2c0bd640ab69865f32ba13786972"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.733849 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.734506 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" podStartSLOduration=7.130989334 podStartE2EDuration="27.734493438s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:13.874286692 +0000 UTC m=+899.230630717" lastFinishedPulling="2025-10-02 09:42:34.477790796 +0000 UTC m=+919.834134821" observedRunningTime="2025-10-02 09:42:39.731505021 +0000 UTC m=+925.087849056" watchObservedRunningTime="2025-10-02 09:42:39.734493438 +0000 UTC m=+925.090837463" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.740272 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" event={"ID":"c7a8160d-c732-4736-a1b0-57999db04976","Type":"ContainerStarted","Data":"38a143308bd37125d95fc02f7e815b8fa6585dc7b1e914213b228d610998ab5c"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.740602 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.742789 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" event={"ID":"1b7f13af-0af5-4ffd-b677-e583068cd598","Type":"ContainerStarted","Data":"ba894db94c02703f830bbd7360e672f429eeb2aebeac4868221fdfc2ff37536a"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.742932 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.746803 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" event={"ID":"3bb0cc56-b9e6-42fe-9122-929e1f7a60e8","Type":"ContainerStarted","Data":"e80c05ca6bb854c43b951b6b9a73c31b36fa6c897fbd709848798888ba190e96"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.747018 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.748914 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" event={"ID":"9c08e6d0-af1a-43cc-a898-d345eba038bb","Type":"ContainerStarted","Data":"947540a0614c57bb345d4b597aa5de968ca316a16bf1a7927f72de058f6af111"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.749097 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.751509 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" event={"ID":"2279a6b9-cecc-4b8b-ac52-041a6abe4b9a","Type":"ContainerStarted","Data":"bba637c40e0d7da796dbc321d3b37c70dd1fc22f49ecc29cd45e8bf8134ed633"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.753614 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" event={"ID":"107fac49-7084-4524-848e-4ebc95387af7","Type":"ContainerStarted","Data":"ae4a04057d2b6d20da7789992d9d55a4ed22a8c440381062b132670ba8561d70"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.755687 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.756146 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" podStartSLOduration=5.136402964 podStartE2EDuration="27.756130508s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.037078781 +0000 UTC m=+899.393422796" lastFinishedPulling="2025-10-02 09:42:36.656806315 +0000 UTC m=+922.013150340" observedRunningTime="2025-10-02 09:42:39.755929212 +0000 UTC m=+925.112273267" watchObservedRunningTime="2025-10-02 09:42:39.756130508 +0000 UTC m=+925.112474543" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.760397 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" event={"ID":"099bc72a-55f2-438f-b90c-5e54cc4e7f8e","Type":"ContainerStarted","Data":"f83a3540154b5dbbe5c3bcf74af06d15669f0a866f5386a5ab0915d9995e9f1a"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.760748 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.767755 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" event={"ID":"bf4ed5a3-c0a1-4b98-837b-20de28931ba8","Type":"ContainerStarted","Data":"79a71d209ee6054ef4aba9bdd7aebc7d2256d6d511e2ab50fab5122809f079a5"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.768516 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.772468 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" event={"ID":"bb9ed734-53de-48a2-ac4d-57c513e74c83","Type":"ContainerStarted","Data":"23155d9b87adf7c98abfa94efe4f750e7451f6c16aa88fb3d72293370389ca92"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.772629 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.776629 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" podStartSLOduration=6.650152714 podStartE2EDuration="27.776610094s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:13.869115691 +0000 UTC m=+899.225459726" lastFinishedPulling="2025-10-02 09:42:34.995573081 +0000 UTC m=+920.351917106" observedRunningTime="2025-10-02 09:42:39.773811382 +0000 UTC m=+925.130155407" watchObservedRunningTime="2025-10-02 09:42:39.776610094 +0000 UTC m=+925.132954119" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.776897 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" event={"ID":"219d8519-1035-4127-a7b3-d59730e87ba8","Type":"ContainerStarted","Data":"f1ec1a0d667bc3715de3b33646144fccb512d4e59e2fb50c8a367384cefaedb4"} Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.777801 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.824303 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" podStartSLOduration=5.276368579 podStartE2EDuration="27.824277252s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.048408101 +0000 UTC m=+899.404752126" lastFinishedPulling="2025-10-02 09:42:36.596316774 +0000 UTC m=+921.952660799" observedRunningTime="2025-10-02 09:42:39.818553215 +0000 UTC m=+925.174897240" watchObservedRunningTime="2025-10-02 09:42:39.824277252 +0000 UTC m=+925.180621287" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.861345 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" podStartSLOduration=5.799735366 podStartE2EDuration="27.86132563s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.60410689 +0000 UTC m=+899.960450915" lastFinishedPulling="2025-10-02 09:42:36.665697154 +0000 UTC m=+922.022041179" observedRunningTime="2025-10-02 09:42:39.857523999 +0000 UTC m=+925.213868024" watchObservedRunningTime="2025-10-02 09:42:39.86132563 +0000 UTC m=+925.217669655" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.862863 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" podStartSLOduration=5.794791283 podStartE2EDuration="27.862854265s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.59070949 +0000 UTC m=+899.947053515" lastFinishedPulling="2025-10-02 09:42:36.658772472 +0000 UTC m=+922.015116497" observedRunningTime="2025-10-02 09:42:39.844013196 +0000 UTC m=+925.200357231" watchObservedRunningTime="2025-10-02 09:42:39.862854265 +0000 UTC m=+925.219198290" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.874746 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" podStartSLOduration=5.857989542 podStartE2EDuration="27.87472581s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.661690736 +0000 UTC m=+900.018034761" lastFinishedPulling="2025-10-02 09:42:36.678427004 +0000 UTC m=+922.034771029" observedRunningTime="2025-10-02 09:42:39.870208089 +0000 UTC m=+925.226552134" watchObservedRunningTime="2025-10-02 09:42:39.87472581 +0000 UTC m=+925.231069825" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.891281 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" podStartSLOduration=5.760129743 podStartE2EDuration="27.891260952s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.467780581 +0000 UTC m=+899.824124606" lastFinishedPulling="2025-10-02 09:42:36.59891179 +0000 UTC m=+921.955255815" observedRunningTime="2025-10-02 09:42:39.889688656 +0000 UTC m=+925.246032691" watchObservedRunningTime="2025-10-02 09:42:39.891260952 +0000 UTC m=+925.247604977" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.912386 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" podStartSLOduration=3.651499892 podStartE2EDuration="27.912366986s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:13.782019435 +0000 UTC m=+899.138363460" lastFinishedPulling="2025-10-02 09:42:38.042886529 +0000 UTC m=+923.399230554" observedRunningTime="2025-10-02 09:42:39.910383379 +0000 UTC m=+925.266727414" watchObservedRunningTime="2025-10-02 09:42:39.912366986 +0000 UTC m=+925.268711011" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.935181 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" podStartSLOduration=6.241126267 podStartE2EDuration="27.93515702s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:13.301576499 +0000 UTC m=+898.657920524" lastFinishedPulling="2025-10-02 09:42:34.995607252 +0000 UTC m=+920.351951277" observedRunningTime="2025-10-02 09:42:39.931411811 +0000 UTC m=+925.287755846" watchObservedRunningTime="2025-10-02 09:42:39.93515702 +0000 UTC m=+925.291501045" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.948740 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" podStartSLOduration=5.386754583 podStartE2EDuration="27.948721515s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.036895466 +0000 UTC m=+899.393239491" lastFinishedPulling="2025-10-02 09:42:36.598862408 +0000 UTC m=+921.955206423" observedRunningTime="2025-10-02 09:42:39.947738926 +0000 UTC m=+925.304082971" watchObservedRunningTime="2025-10-02 09:42:39.948721515 +0000 UTC m=+925.305065540" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.964891 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" podStartSLOduration=4.824778582 podStartE2EDuration="27.964874125s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:13.458826177 +0000 UTC m=+898.815170202" lastFinishedPulling="2025-10-02 09:42:36.59892173 +0000 UTC m=+921.955265745" observedRunningTime="2025-10-02 09:42:39.962395253 +0000 UTC m=+925.318739278" watchObservedRunningTime="2025-10-02 09:42:39.964874125 +0000 UTC m=+925.321218150" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.997380 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" podStartSLOduration=10.881427812 podStartE2EDuration="27.997358131s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:13.883729086 +0000 UTC m=+899.240073111" lastFinishedPulling="2025-10-02 09:42:30.999659405 +0000 UTC m=+916.356003430" observedRunningTime="2025-10-02 09:42:39.981016045 +0000 UTC m=+925.337360090" watchObservedRunningTime="2025-10-02 09:42:39.997358131 +0000 UTC m=+925.353702156" Oct 02 09:42:39 crc kubenswrapper[5035]: I1002 09:42:39.998176 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" podStartSLOduration=7.582487908 podStartE2EDuration="27.998170194s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:13.682736425 +0000 UTC m=+899.039080450" lastFinishedPulling="2025-10-02 09:42:34.098418721 +0000 UTC m=+919.454762736" observedRunningTime="2025-10-02 09:42:39.993777576 +0000 UTC m=+925.350121601" watchObservedRunningTime="2025-10-02 09:42:39.998170194 +0000 UTC m=+925.354514219" Oct 02 09:42:40 crc kubenswrapper[5035]: I1002 09:42:40.037657 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" podStartSLOduration=5.991154889 podStartE2EDuration="28.037637093s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.554247178 +0000 UTC m=+899.910591203" lastFinishedPulling="2025-10-02 09:42:36.600729382 +0000 UTC m=+921.957073407" observedRunningTime="2025-10-02 09:42:40.021933616 +0000 UTC m=+925.378277641" watchObservedRunningTime="2025-10-02 09:42:40.037637093 +0000 UTC m=+925.393981108" Oct 02 09:42:40 crc kubenswrapper[5035]: I1002 09:42:40.041141 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" podStartSLOduration=6.180930984 podStartE2EDuration="28.041127295s" podCreationTimestamp="2025-10-02 09:42:12 +0000 UTC" firstStartedPulling="2025-10-02 09:42:14.738615986 +0000 UTC m=+900.094960011" lastFinishedPulling="2025-10-02 09:42:36.598812287 +0000 UTC m=+921.955156322" observedRunningTime="2025-10-02 09:42:40.037370866 +0000 UTC m=+925.393714891" watchObservedRunningTime="2025-10-02 09:42:40.041127295 +0000 UTC m=+925.397471320" Oct 02 09:42:40 crc kubenswrapper[5035]: I1002 09:42:40.787672 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqrs8" event={"ID":"8b418acc-8124-4b14-bb79-c036405f371b","Type":"ContainerStarted","Data":"a04abc63af5caf23a7033981324af7260c43ba3a8251a6567fbfbebda3dffc19"} Oct 02 09:42:40 crc kubenswrapper[5035]: I1002 09:42:40.790060 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:40 crc kubenswrapper[5035]: I1002 09:42:40.812930 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zqrs8" podStartSLOduration=18.076152408 podStartE2EDuration="20.812905975s" podCreationTimestamp="2025-10-02 09:42:20 +0000 UTC" firstStartedPulling="2025-10-02 09:42:37.618764781 +0000 UTC m=+922.975108806" lastFinishedPulling="2025-10-02 09:42:40.355518348 +0000 UTC m=+925.711862373" observedRunningTime="2025-10-02 09:42:40.807742854 +0000 UTC m=+926.164086879" watchObservedRunningTime="2025-10-02 09:42:40.812905975 +0000 UTC m=+926.169250000" Oct 02 09:42:41 crc kubenswrapper[5035]: I1002 09:42:41.231026 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:41 crc kubenswrapper[5035]: I1002 09:42:41.231105 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.270271 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zqrs8" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="registry-server" probeResult="failure" output=< Oct 02 09:42:42 crc kubenswrapper[5035]: timeout: failed to connect service ":50051" within 1s Oct 02 09:42:42 crc kubenswrapper[5035]: > Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.414439 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-qmrqc" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.538808 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-5znkc" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.705434 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-hp45m" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.715158 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-msx6h" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.725049 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-sb4zr" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.756199 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-p72mn" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.826835 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-dtzbf" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.923476 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-zh2nk" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.933354 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-bcnlz" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.967211 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-sgg6c" Oct 02 09:42:42 crc kubenswrapper[5035]: I1002 09:42:42.989904 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-nsrkb" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.068646 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-rs2lh" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.092378 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-8f4pt" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.137674 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-2lqs2" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.146069 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-49mnw" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.161363 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-6pm9g" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.229788 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-msfdb" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.275503 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-hvhzg" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.286220 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-m58vc" Oct 02 09:42:43 crc kubenswrapper[5035]: I1002 09:42:43.628633 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9grgw" Oct 02 09:42:51 crc kubenswrapper[5035]: I1002 09:42:51.271013 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:51 crc kubenswrapper[5035]: I1002 09:42:51.313472 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:51 crc kubenswrapper[5035]: I1002 09:42:51.512498 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqrs8"] Oct 02 09:42:52 crc kubenswrapper[5035]: I1002 09:42:52.691971 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-d87866488-xc6dw" Oct 02 09:42:52 crc kubenswrapper[5035]: I1002 09:42:52.871894 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zqrs8" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="registry-server" containerID="cri-o://a04abc63af5caf23a7033981324af7260c43ba3a8251a6567fbfbebda3dffc19" gracePeriod=2 Oct 02 09:42:53 crc kubenswrapper[5035]: I1002 09:42:53.881339 5035 generic.go:334] "Generic (PLEG): container finished" podID="8b418acc-8124-4b14-bb79-c036405f371b" containerID="a04abc63af5caf23a7033981324af7260c43ba3a8251a6567fbfbebda3dffc19" exitCode=0 Oct 02 09:42:53 crc kubenswrapper[5035]: I1002 09:42:53.881383 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqrs8" event={"ID":"8b418acc-8124-4b14-bb79-c036405f371b","Type":"ContainerDied","Data":"a04abc63af5caf23a7033981324af7260c43ba3a8251a6567fbfbebda3dffc19"} Oct 02 09:42:55 crc kubenswrapper[5035]: I1002 09:42:55.538278 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:42:55 crc kubenswrapper[5035]: I1002 09:42:55.538656 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.327295 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.423210 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcrl2\" (UniqueName: \"kubernetes.io/projected/8b418acc-8124-4b14-bb79-c036405f371b-kube-api-access-rcrl2\") pod \"8b418acc-8124-4b14-bb79-c036405f371b\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.423267 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-utilities\") pod \"8b418acc-8124-4b14-bb79-c036405f371b\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.423361 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-catalog-content\") pod \"8b418acc-8124-4b14-bb79-c036405f371b\" (UID: \"8b418acc-8124-4b14-bb79-c036405f371b\") " Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.424469 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-utilities" (OuterVolumeSpecName: "utilities") pod "8b418acc-8124-4b14-bb79-c036405f371b" (UID: "8b418acc-8124-4b14-bb79-c036405f371b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.429758 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b418acc-8124-4b14-bb79-c036405f371b-kube-api-access-rcrl2" (OuterVolumeSpecName: "kube-api-access-rcrl2") pod "8b418acc-8124-4b14-bb79-c036405f371b" (UID: "8b418acc-8124-4b14-bb79-c036405f371b"). InnerVolumeSpecName "kube-api-access-rcrl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.466747 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b418acc-8124-4b14-bb79-c036405f371b" (UID: "8b418acc-8124-4b14-bb79-c036405f371b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.524921 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.524979 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcrl2\" (UniqueName: \"kubernetes.io/projected/8b418acc-8124-4b14-bb79-c036405f371b-kube-api-access-rcrl2\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.525082 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b418acc-8124-4b14-bb79-c036405f371b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.926227 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zqrs8" event={"ID":"8b418acc-8124-4b14-bb79-c036405f371b","Type":"ContainerDied","Data":"08e0f01aced94efaf99196b3030eec1ce2d99a43d4bd2864dd4d39234575295f"} Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.926276 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zqrs8" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.926316 5035 scope.go:117] "RemoveContainer" containerID="a04abc63af5caf23a7033981324af7260c43ba3a8251a6567fbfbebda3dffc19" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.950138 5035 scope.go:117] "RemoveContainer" containerID="9d6a12fe60aee60d90970d6f950e819744b7e5b86d06affbb815a2cd7e6930e7" Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.958709 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zqrs8"] Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.966189 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zqrs8"] Oct 02 09:42:59 crc kubenswrapper[5035]: I1002 09:42:59.989594 5035 scope.go:117] "RemoveContainer" containerID="71e156ce725f28679f9d21215941390f217f47fa3858c8c275c62dfade25315e" Oct 02 09:43:00 crc kubenswrapper[5035]: I1002 09:43:00.171774 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b418acc-8124-4b14-bb79-c036405f371b" path="/var/lib/kubelet/pods/8b418acc-8124-4b14-bb79-c036405f371b/volumes" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.054411 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xz8qb"] Oct 02 09:43:10 crc kubenswrapper[5035]: E1002 09:43:10.055783 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="extract-content" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.055804 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="extract-content" Oct 02 09:43:10 crc kubenswrapper[5035]: E1002 09:43:10.055850 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="extract-utilities" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.055859 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="extract-utilities" Oct 02 09:43:10 crc kubenswrapper[5035]: E1002 09:43:10.055885 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="registry-server" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.055892 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="registry-server" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.056205 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b418acc-8124-4b14-bb79-c036405f371b" containerName="registry-server" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.057049 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.059986 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.060171 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.060362 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.061287 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-znfgz" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.063294 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xz8qb"] Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.133118 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2bmg9"] Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.134395 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.139202 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.147034 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2bmg9"] Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.191519 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-config\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.191619 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldpgs\" (UniqueName: \"kubernetes.io/projected/c8b9019d-0c99-4070-88d1-156f57617e20-kube-api-access-ldpgs\") pod \"dnsmasq-dns-675f4bcbfc-xz8qb\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.191662 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.191712 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9svz\" (UniqueName: \"kubernetes.io/projected/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-kube-api-access-g9svz\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.191765 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8b9019d-0c99-4070-88d1-156f57617e20-config\") pod \"dnsmasq-dns-675f4bcbfc-xz8qb\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.294040 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9svz\" (UniqueName: \"kubernetes.io/projected/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-kube-api-access-g9svz\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.294146 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8b9019d-0c99-4070-88d1-156f57617e20-config\") pod \"dnsmasq-dns-675f4bcbfc-xz8qb\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.294189 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-config\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.294275 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldpgs\" (UniqueName: \"kubernetes.io/projected/c8b9019d-0c99-4070-88d1-156f57617e20-kube-api-access-ldpgs\") pod \"dnsmasq-dns-675f4bcbfc-xz8qb\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.294339 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.296213 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8b9019d-0c99-4070-88d1-156f57617e20-config\") pod \"dnsmasq-dns-675f4bcbfc-xz8qb\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.296408 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.296458 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-config\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.317448 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9svz\" (UniqueName: \"kubernetes.io/projected/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-kube-api-access-g9svz\") pod \"dnsmasq-dns-78dd6ddcc-2bmg9\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.317970 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldpgs\" (UniqueName: \"kubernetes.io/projected/c8b9019d-0c99-4070-88d1-156f57617e20-kube-api-access-ldpgs\") pod \"dnsmasq-dns-675f4bcbfc-xz8qb\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.383942 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.453434 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.719665 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2bmg9"] Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.725734 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.896653 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xz8qb"] Oct 02 09:43:10 crc kubenswrapper[5035]: W1002 09:43:10.899263 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8b9019d_0c99_4070_88d1_156f57617e20.slice/crio-20392c9e11acab55c3c7dada9c8d6ce8d7b60f4b28a5c5eac8806d7c0c6d3f8f WatchSource:0}: Error finding container 20392c9e11acab55c3c7dada9c8d6ce8d7b60f4b28a5c5eac8806d7c0c6d3f8f: Status 404 returned error can't find the container with id 20392c9e11acab55c3c7dada9c8d6ce8d7b60f4b28a5c5eac8806d7c0c6d3f8f Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.997294 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" event={"ID":"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2","Type":"ContainerStarted","Data":"ab27d08272debcd11e7d171b8a6c84247b5a1da55920b9791137cb852d42ca32"} Oct 02 09:43:10 crc kubenswrapper[5035]: I1002 09:43:10.998495 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" event={"ID":"c8b9019d-0c99-4070-88d1-156f57617e20","Type":"ContainerStarted","Data":"20392c9e11acab55c3c7dada9c8d6ce8d7b60f4b28a5c5eac8806d7c0c6d3f8f"} Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.295549 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xz8qb"] Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.326727 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2ttrv"] Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.330548 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.339195 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2ttrv"] Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.446995 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qflg\" (UniqueName: \"kubernetes.io/projected/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-kube-api-access-7qflg\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.447086 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-config\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.447139 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.549981 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.550055 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qflg\" (UniqueName: \"kubernetes.io/projected/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-kube-api-access-7qflg\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.550250 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-config\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.552339 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-config\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.554314 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.584123 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qflg\" (UniqueName: \"kubernetes.io/projected/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-kube-api-access-7qflg\") pod \"dnsmasq-dns-5ccc8479f9-2ttrv\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.588308 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2bmg9"] Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.609212 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9bs5d"] Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.610443 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.633299 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9bs5d"] Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.659672 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.758786 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.758873 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-config\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.758903 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7krhs\" (UniqueName: \"kubernetes.io/projected/59c56020-0f32-4362-8ce8-a66ed3d219b4-kube-api-access-7krhs\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.859747 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.859801 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-config\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.859823 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7krhs\" (UniqueName: \"kubernetes.io/projected/59c56020-0f32-4362-8ce8-a66ed3d219b4-kube-api-access-7krhs\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.861041 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.863297 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-config\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.882758 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7krhs\" (UniqueName: \"kubernetes.io/projected/59c56020-0f32-4362-8ce8-a66ed3d219b4-kube-api-access-7krhs\") pod \"dnsmasq-dns-57d769cc4f-9bs5d\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:12 crc kubenswrapper[5035]: I1002 09:43:12.943392 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.191366 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2ttrv"] Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.428333 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9bs5d"] Oct 02 09:43:13 crc kubenswrapper[5035]: W1002 09:43:13.447499 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59c56020_0f32_4362_8ce8_a66ed3d219b4.slice/crio-bce56cfed1b215af4b1f4b1d5a6eb601b6f75651ab0dee0ad01cda423e484008 WatchSource:0}: Error finding container bce56cfed1b215af4b1f4b1d5a6eb601b6f75651ab0dee0ad01cda423e484008: Status 404 returned error can't find the container with id bce56cfed1b215af4b1f4b1d5a6eb601b6f75651ab0dee0ad01cda423e484008 Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.468701 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.470323 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.473266 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-pfxfj" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.473356 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.473621 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.474035 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.474067 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.477892 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.478107 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.482702 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.584354 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d7eafce-151f-43f3-8f74-cbfc9f368643-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.584403 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.584609 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.584950 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkc96\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-kube-api-access-lkc96\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.585091 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.585292 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.585361 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.585390 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d7eafce-151f-43f3-8f74-cbfc9f368643-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.585496 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.585645 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.585676 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687089 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687174 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687222 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d7eafce-151f-43f3-8f74-cbfc9f368643-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687245 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687409 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687429 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687512 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d7eafce-151f-43f3-8f74-cbfc9f368643-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687568 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687590 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687611 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkc96\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-kube-api-access-lkc96\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.687641 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.688127 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.688132 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.688484 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.688903 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.689924 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.698841 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.698972 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d7eafce-151f-43f3-8f74-cbfc9f368643-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.704315 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d7eafce-151f-43f3-8f74-cbfc9f368643-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.705494 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.706366 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d7eafce-151f-43f3-8f74-cbfc9f368643-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.714728 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.717296 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkc96\" (UniqueName: \"kubernetes.io/projected/3d7eafce-151f-43f3-8f74-cbfc9f368643-kube-api-access-lkc96\") pod \"rabbitmq-cell1-server-0\" (UID: \"3d7eafce-151f-43f3-8f74-cbfc9f368643\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.745481 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.762140 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.766892 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.768691 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.770888 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.771069 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.771240 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.771510 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.771514 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.772647 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-v845b" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.808917 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.891808 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.891885 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.891952 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.891978 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d93ca48d-a2a5-4036-930d-948d935cce98-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.892025 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.892062 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.892089 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d93ca48d-a2a5-4036-930d-948d935cce98-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.892129 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.892157 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwdrn\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-kube-api-access-cwdrn\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.892185 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.892216 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-config-data\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993203 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993445 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d93ca48d-a2a5-4036-930d-948d935cce98-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993479 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993497 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwdrn\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-kube-api-access-cwdrn\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993517 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993569 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-config-data\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993586 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993612 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993650 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993666 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d93ca48d-a2a5-4036-930d-948d935cce98-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993693 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.993921 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.995284 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.996095 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.996891 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-config-data\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.997016 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:13 crc kubenswrapper[5035]: I1002 09:43:13.997950 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d93ca48d-a2a5-4036-930d-948d935cce98-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.008570 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.013235 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d93ca48d-a2a5-4036-930d-948d935cce98-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.015656 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d93ca48d-a2a5-4036-930d-948d935cce98-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.016856 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.034626 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.048892 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwdrn\" (UniqueName: \"kubernetes.io/projected/d93ca48d-a2a5-4036-930d-948d935cce98-kube-api-access-cwdrn\") pod \"rabbitmq-server-0\" (UID: \"d93ca48d-a2a5-4036-930d-948d935cce98\") " pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.056692 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" event={"ID":"59c56020-0f32-4362-8ce8-a66ed3d219b4","Type":"ContainerStarted","Data":"bce56cfed1b215af4b1f4b1d5a6eb601b6f75651ab0dee0ad01cda423e484008"} Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.089138 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" event={"ID":"c4138bce-f108-4a2c-b2fa-8a44f264ccc7","Type":"ContainerStarted","Data":"624b1a942bfc1508fd5194c9695c9db1e356274572c62dae0757cc07d8877b60"} Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.096838 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.225726 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 09:43:14 crc kubenswrapper[5035]: W1002 09:43:14.246339 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d7eafce_151f_43f3_8f74_cbfc9f368643.slice/crio-6781d2dc9b9c3abc6cd990c38c9185fc769048375b523aa5edeab07fa39adc01 WatchSource:0}: Error finding container 6781d2dc9b9c3abc6cd990c38c9185fc769048375b523aa5edeab07fa39adc01: Status 404 returned error can't find the container with id 6781d2dc9b9c3abc6cd990c38c9185fc769048375b523aa5edeab07fa39adc01 Oct 02 09:43:14 crc kubenswrapper[5035]: I1002 09:43:14.711739 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 09:43:14 crc kubenswrapper[5035]: W1002 09:43:14.719219 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd93ca48d_a2a5_4036_930d_948d935cce98.slice/crio-4708379b5f7118d7f6ec7f25d56fe86aaaacc98ebe609f47512dc070590f8b38 WatchSource:0}: Error finding container 4708379b5f7118d7f6ec7f25d56fe86aaaacc98ebe609f47512dc070590f8b38: Status 404 returned error can't find the container with id 4708379b5f7118d7f6ec7f25d56fe86aaaacc98ebe609f47512dc070590f8b38 Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.155031 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d93ca48d-a2a5-4036-930d-948d935cce98","Type":"ContainerStarted","Data":"4708379b5f7118d7f6ec7f25d56fe86aaaacc98ebe609f47512dc070590f8b38"} Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.162949 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3d7eafce-151f-43f3-8f74-cbfc9f368643","Type":"ContainerStarted","Data":"6781d2dc9b9c3abc6cd990c38c9185fc769048375b523aa5edeab07fa39adc01"} Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.295581 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.297026 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.305290 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.307448 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.307640 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.307451 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.307782 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5r7dk" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.309838 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.338899 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421149 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421289 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/da088dc0-8011-44e7-973b-66e75f33549c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421374 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421563 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421629 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-config-data-default\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421745 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421812 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flcrd\" (UniqueName: \"kubernetes.io/projected/da088dc0-8011-44e7-973b-66e75f33549c-kube-api-access-flcrd\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.421970 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-secrets\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.422038 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-kolla-config\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523292 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flcrd\" (UniqueName: \"kubernetes.io/projected/da088dc0-8011-44e7-973b-66e75f33549c-kube-api-access-flcrd\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523359 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-secrets\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523397 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-kolla-config\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523450 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523477 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/da088dc0-8011-44e7-973b-66e75f33549c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523545 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523593 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523624 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-config-data-default\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.523662 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.524016 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/da088dc0-8011-44e7-973b-66e75f33549c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.524260 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.524891 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-config-data-default\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.525052 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-kolla-config\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.525204 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da088dc0-8011-44e7-973b-66e75f33549c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.530388 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-secrets\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.530401 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.532729 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da088dc0-8011-44e7-973b-66e75f33549c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.542668 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flcrd\" (UniqueName: \"kubernetes.io/projected/da088dc0-8011-44e7-973b-66e75f33549c-kube-api-access-flcrd\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.550644 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"da088dc0-8011-44e7-973b-66e75f33549c\") " pod="openstack/openstack-galera-0" Oct 02 09:43:15 crc kubenswrapper[5035]: I1002 09:43:15.642687 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.663149 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.668669 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.672473 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.672916 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.673252 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-fk5cx" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.673554 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.677037 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.768871 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.768943 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.768964 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbc49\" (UniqueName: \"kubernetes.io/projected/9f58a85f-3d79-4ed1-a856-c491969c730f-kube-api-access-fbc49\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.769034 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9f58a85f-3d79-4ed1-a856-c491969c730f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.769061 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.769091 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.769106 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.769125 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.769307 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.873834 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.873910 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.873956 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.873976 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbc49\" (UniqueName: \"kubernetes.io/projected/9f58a85f-3d79-4ed1-a856-c491969c730f-kube-api-access-fbc49\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.873997 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9f58a85f-3d79-4ed1-a856-c491969c730f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.874013 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.874038 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.874054 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.874070 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.874242 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.875017 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9f58a85f-3d79-4ed1-a856-c491969c730f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.876375 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.877054 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.877191 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9f58a85f-3d79-4ed1-a856-c491969c730f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.880691 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.886600 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.901025 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.908114 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbc49\" (UniqueName: \"kubernetes.io/projected/9f58a85f-3d79-4ed1-a856-c491969c730f-kube-api-access-fbc49\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.918618 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/9f58a85f-3d79-4ed1-a856-c491969c730f-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"9f58a85f-3d79-4ed1-a856-c491969c730f\") " pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.958333 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.964742 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.969393 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.970326 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.970450 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-rgj7w" Oct 02 09:43:16 crc kubenswrapper[5035]: I1002 09:43:16.979189 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:16.998957 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.078296 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b08491fc-549c-4176-a8e8-5f733d838e3d-config-data\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.078360 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08491fc-549c-4176-a8e8-5f733d838e3d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.078458 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k48r2\" (UniqueName: \"kubernetes.io/projected/b08491fc-549c-4176-a8e8-5f733d838e3d-kube-api-access-k48r2\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.078687 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08491fc-549c-4176-a8e8-5f733d838e3d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.078755 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b08491fc-549c-4176-a8e8-5f733d838e3d-kolla-config\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.180060 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08491fc-549c-4176-a8e8-5f733d838e3d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.180110 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b08491fc-549c-4176-a8e8-5f733d838e3d-kolla-config\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.180152 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b08491fc-549c-4176-a8e8-5f733d838e3d-config-data\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.180175 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08491fc-549c-4176-a8e8-5f733d838e3d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.180218 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k48r2\" (UniqueName: \"kubernetes.io/projected/b08491fc-549c-4176-a8e8-5f733d838e3d-kube-api-access-k48r2\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.181827 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b08491fc-549c-4176-a8e8-5f733d838e3d-config-data\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.182016 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b08491fc-549c-4176-a8e8-5f733d838e3d-kolla-config\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.191178 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b08491fc-549c-4176-a8e8-5f733d838e3d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.203121 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b08491fc-549c-4176-a8e8-5f733d838e3d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.205916 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k48r2\" (UniqueName: \"kubernetes.io/projected/b08491fc-549c-4176-a8e8-5f733d838e3d-kube-api-access-k48r2\") pod \"memcached-0\" (UID: \"b08491fc-549c-4176-a8e8-5f733d838e3d\") " pod="openstack/memcached-0" Oct 02 09:43:17 crc kubenswrapper[5035]: I1002 09:43:17.315265 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 09:43:18 crc kubenswrapper[5035]: I1002 09:43:18.971375 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:43:18 crc kubenswrapper[5035]: I1002 09:43:18.972493 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:43:18 crc kubenswrapper[5035]: I1002 09:43:18.974930 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-btqtx" Oct 02 09:43:18 crc kubenswrapper[5035]: I1002 09:43:18.985301 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:43:19 crc kubenswrapper[5035]: I1002 09:43:19.015703 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkfmf\" (UniqueName: \"kubernetes.io/projected/f677fa29-64c1-45f7-8f6d-c4c15f83db60-kube-api-access-kkfmf\") pod \"kube-state-metrics-0\" (UID: \"f677fa29-64c1-45f7-8f6d-c4c15f83db60\") " pod="openstack/kube-state-metrics-0" Oct 02 09:43:19 crc kubenswrapper[5035]: I1002 09:43:19.117268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkfmf\" (UniqueName: \"kubernetes.io/projected/f677fa29-64c1-45f7-8f6d-c4c15f83db60-kube-api-access-kkfmf\") pod \"kube-state-metrics-0\" (UID: \"f677fa29-64c1-45f7-8f6d-c4c15f83db60\") " pod="openstack/kube-state-metrics-0" Oct 02 09:43:19 crc kubenswrapper[5035]: I1002 09:43:19.148168 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkfmf\" (UniqueName: \"kubernetes.io/projected/f677fa29-64c1-45f7-8f6d-c4c15f83db60-kube-api-access-kkfmf\") pod \"kube-state-metrics-0\" (UID: \"f677fa29-64c1-45f7-8f6d-c4c15f83db60\") " pod="openstack/kube-state-metrics-0" Oct 02 09:43:19 crc kubenswrapper[5035]: I1002 09:43:19.292029 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.006660 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jvdts"] Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.013136 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-r5m97"] Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.013508 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.015259 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.016928 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-p2qtj" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.017133 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.017176 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jvdts"] Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.017290 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.043313 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-r5m97"] Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063433 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-run\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063480 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-ovn-controller-tls-certs\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063505 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-etc-ovs\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063542 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpz99\" (UniqueName: \"kubernetes.io/projected/89658bb0-5353-4e67-95cd-cdb040086903-kube-api-access-mpz99\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063568 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-run\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063583 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-log\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063616 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-log-ovn\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063649 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89658bb0-5353-4e67-95cd-cdb040086903-scripts\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063672 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-run-ovn\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063694 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-scripts\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063708 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-combined-ca-bundle\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063737 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs6m8\" (UniqueName: \"kubernetes.io/projected/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-kube-api-access-rs6m8\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.063764 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-lib\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170109 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-ovn-controller-tls-certs\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170251 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-etc-ovs\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170307 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpz99\" (UniqueName: \"kubernetes.io/projected/89658bb0-5353-4e67-95cd-cdb040086903-kube-api-access-mpz99\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170378 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-run\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170405 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-log\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170450 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-log-ovn\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170567 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89658bb0-5353-4e67-95cd-cdb040086903-scripts\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170623 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-run-ovn\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170668 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-scripts\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170701 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-combined-ca-bundle\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170781 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs6m8\" (UniqueName: \"kubernetes.io/projected/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-kube-api-access-rs6m8\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170858 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-lib\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170895 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-run\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.170901 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-log\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.171032 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-run-ovn\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.171181 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-log-ovn\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.171239 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-run\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.171745 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-var-lib\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.171823 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-var-run\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.173600 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-scripts\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.173785 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/89658bb0-5353-4e67-95cd-cdb040086903-scripts\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.174118 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/89658bb0-5353-4e67-95cd-cdb040086903-etc-ovs\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.176802 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-ovn-controller-tls-certs\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.177986 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-combined-ca-bundle\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.194699 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpz99\" (UniqueName: \"kubernetes.io/projected/89658bb0-5353-4e67-95cd-cdb040086903-kube-api-access-mpz99\") pod \"ovn-controller-ovs-r5m97\" (UID: \"89658bb0-5353-4e67-95cd-cdb040086903\") " pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.196098 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs6m8\" (UniqueName: \"kubernetes.io/projected/57ea64fd-e931-45ab-9f0e-678c3c9ba7a2-kube-api-access-rs6m8\") pod \"ovn-controller-jvdts\" (UID: \"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2\") " pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.290624 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.292020 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.299796 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.299954 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.299996 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.300126 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.300384 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rttrh" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.347686 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.358128 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.371409 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.372583 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07a982ff-c270-4133-a1a0-612dc7317d4b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.372642 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.372665 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.372689 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.372849 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a982ff-c270-4133-a1a0-612dc7317d4b-config\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.372963 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.373082 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07a982ff-c270-4133-a1a0-612dc7317d4b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.373278 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-876n5\" (UniqueName: \"kubernetes.io/projected/07a982ff-c270-4133-a1a0-612dc7317d4b-kube-api-access-876n5\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475213 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a982ff-c270-4133-a1a0-612dc7317d4b-config\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475294 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475321 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07a982ff-c270-4133-a1a0-612dc7317d4b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475381 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-876n5\" (UniqueName: \"kubernetes.io/projected/07a982ff-c270-4133-a1a0-612dc7317d4b-kube-api-access-876n5\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475425 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07a982ff-c270-4133-a1a0-612dc7317d4b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475474 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475496 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475520 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.475890 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.476234 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a982ff-c270-4133-a1a0-612dc7317d4b-config\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.480393 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/07a982ff-c270-4133-a1a0-612dc7317d4b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.486899 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/07a982ff-c270-4133-a1a0-612dc7317d4b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.497175 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.497416 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.499493 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/07a982ff-c270-4133-a1a0-612dc7317d4b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.510384 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-876n5\" (UniqueName: \"kubernetes.io/projected/07a982ff-c270-4133-a1a0-612dc7317d4b-kube-api-access-876n5\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.526681 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-nb-0\" (UID: \"07a982ff-c270-4133-a1a0-612dc7317d4b\") " pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:22 crc kubenswrapper[5035]: I1002 09:43:22.658631 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:25 crc kubenswrapper[5035]: I1002 09:43:25.538503 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:43:25 crc kubenswrapper[5035]: I1002 09:43:25.538958 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.189293 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.191430 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.193396 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.193666 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2sf4t" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.193959 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.203626 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.210061 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.246339 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-config\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.246425 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qcf4\" (UniqueName: \"kubernetes.io/projected/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-kube-api-access-5qcf4\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.246462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.246486 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.246549 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.246887 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.246978 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.247148 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348430 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348478 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348506 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348585 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348646 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348666 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-config\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348693 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qcf4\" (UniqueName: \"kubernetes.io/projected/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-kube-api-access-5qcf4\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.348715 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.349195 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.349442 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.350138 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.350304 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-config\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.355919 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.356243 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.356919 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.365050 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qcf4\" (UniqueName: \"kubernetes.io/projected/0578ddf1-8297-4720-b0c1-8fa7b7dce11d-kube-api-access-5qcf4\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.376575 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"0578ddf1-8297-4720-b0c1-8fa7b7dce11d\") " pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:26 crc kubenswrapper[5035]: I1002 09:43:26.522374 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.194834 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.195434 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lkc96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(3d7eafce-151f-43f3-8f74-cbfc9f368643): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.196595 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="3d7eafce-151f-43f3-8f74-cbfc9f368643" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.288454 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="3d7eafce-151f-43f3-8f74-cbfc9f368643" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.931173 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.931813 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g9svz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-2bmg9_openstack(68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.933011 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" podUID="68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.985133 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.985299 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7krhs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-9bs5d_openstack(59c56020-0f32-4362-8ce8-a66ed3d219b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.986868 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.991259 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.991386 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7qflg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-2ttrv_openstack(c4138bce-f108-4a2c-b2fa-8a44f264ccc7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:43:28 crc kubenswrapper[5035]: E1002 09:43:28.993263 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" Oct 02 09:43:29 crc kubenswrapper[5035]: E1002 09:43:29.049892 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 09:43:29 crc kubenswrapper[5035]: E1002 09:43:29.050026 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ldpgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-xz8qb_openstack(c8b9019d-0c99-4070-88d1-156f57617e20): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:43:29 crc kubenswrapper[5035]: E1002 09:43:29.052121 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" podUID="c8b9019d-0c99-4070-88d1-156f57617e20" Oct 02 09:43:29 crc kubenswrapper[5035]: E1002 09:43:29.298061 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" Oct 02 09:43:29 crc kubenswrapper[5035]: E1002 09:43:29.304432 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.308793 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.432983 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.539754 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.704095 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jvdts"] Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.718872 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.864224 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.872335 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.890671 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-r5m97"] Oct 02 09:43:29 crc kubenswrapper[5035]: W1002 09:43:29.892812 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89658bb0_5353_4e67_95cd_cdb040086903.slice/crio-8d3fe41ca8fcbaad1000219b6b96a563f1a0d9586da77b4fc3c00749e4298700 WatchSource:0}: Error finding container 8d3fe41ca8fcbaad1000219b6b96a563f1a0d9586da77b4fc3c00749e4298700: Status 404 returned error can't find the container with id 8d3fe41ca8fcbaad1000219b6b96a563f1a0d9586da77b4fc3c00749e4298700 Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.916633 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9svz\" (UniqueName: \"kubernetes.io/projected/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-kube-api-access-g9svz\") pod \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.916733 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8b9019d-0c99-4070-88d1-156f57617e20-config\") pod \"c8b9019d-0c99-4070-88d1-156f57617e20\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.916767 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldpgs\" (UniqueName: \"kubernetes.io/projected/c8b9019d-0c99-4070-88d1-156f57617e20-kube-api-access-ldpgs\") pod \"c8b9019d-0c99-4070-88d1-156f57617e20\" (UID: \"c8b9019d-0c99-4070-88d1-156f57617e20\") " Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.916787 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-config\") pod \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.916816 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-dns-svc\") pod \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\" (UID: \"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2\") " Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.917907 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8b9019d-0c99-4070-88d1-156f57617e20-config" (OuterVolumeSpecName: "config") pod "c8b9019d-0c99-4070-88d1-156f57617e20" (UID: "c8b9019d-0c99-4070-88d1-156f57617e20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.917977 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-config" (OuterVolumeSpecName: "config") pod "68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2" (UID: "68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.918032 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2" (UID: "68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.923379 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-kube-api-access-g9svz" (OuterVolumeSpecName: "kube-api-access-g9svz") pod "68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2" (UID: "68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2"). InnerVolumeSpecName "kube-api-access-g9svz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:29 crc kubenswrapper[5035]: I1002 09:43:29.923864 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8b9019d-0c99-4070-88d1-156f57617e20-kube-api-access-ldpgs" (OuterVolumeSpecName: "kube-api-access-ldpgs") pod "c8b9019d-0c99-4070-88d1-156f57617e20" (UID: "c8b9019d-0c99-4070-88d1-156f57617e20"). InnerVolumeSpecName "kube-api-access-ldpgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.018964 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9svz\" (UniqueName: \"kubernetes.io/projected/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-kube-api-access-g9svz\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.018997 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8b9019d-0c99-4070-88d1-156f57617e20-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.019007 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldpgs\" (UniqueName: \"kubernetes.io/projected/c8b9019d-0c99-4070-88d1-156f57617e20-kube-api-access-ldpgs\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.019017 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.019027 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.122589 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 09:43:30 crc kubenswrapper[5035]: W1002 09:43:30.129909 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07a982ff_c270_4133_a1a0_612dc7317d4b.slice/crio-8c664772e96e771fa88b14644ba2105617d4657907a3f5ef26e440b497f9a674 WatchSource:0}: Error finding container 8c664772e96e771fa88b14644ba2105617d4657907a3f5ef26e440b497f9a674: Status 404 returned error can't find the container with id 8c664772e96e771fa88b14644ba2105617d4657907a3f5ef26e440b497f9a674 Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.303548 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d93ca48d-a2a5-4036-930d-948d935cce98","Type":"ContainerStarted","Data":"e36fa1d3fccf3dedee1297a6e22e93d2b045eb47fc9e1ce5249d3fb5da8dd15e"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.307221 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"da088dc0-8011-44e7-973b-66e75f33549c","Type":"ContainerStarted","Data":"068e4712ecf7db0b94ccf0d9939f02603bb8a031865a6d935acdf4e162217feb"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.318936 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9f58a85f-3d79-4ed1-a856-c491969c730f","Type":"ContainerStarted","Data":"14599a4cb2b2ba3d466501c770dfee510c502f1c93d8f8024651880051e34873"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.320655 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts" event={"ID":"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2","Type":"ContainerStarted","Data":"e9e33fc35910b4bd5cab341e04c8ee2d5c9b13fca361e41c4e147eb30346c7a8"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.322912 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b08491fc-549c-4176-a8e8-5f733d838e3d","Type":"ContainerStarted","Data":"c7f0f623703b7273d5131f7c032a7f9bfdc9a8d14462dc835dec07ed2069a9bb"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.335524 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f677fa29-64c1-45f7-8f6d-c4c15f83db60","Type":"ContainerStarted","Data":"ffa09eef49ab5dff9dcf76901125979935e5a30ff6757fc7afc1a38c4a3eb629"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.338202 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.338207 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2bmg9" event={"ID":"68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2","Type":"ContainerDied","Data":"ab27d08272debcd11e7d171b8a6c84247b5a1da55920b9791137cb852d42ca32"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.340867 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"07a982ff-c270-4133-a1a0-612dc7317d4b","Type":"ContainerStarted","Data":"8c664772e96e771fa88b14644ba2105617d4657907a3f5ef26e440b497f9a674"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.344213 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" event={"ID":"c8b9019d-0c99-4070-88d1-156f57617e20","Type":"ContainerDied","Data":"20392c9e11acab55c3c7dada9c8d6ce8d7b60f4b28a5c5eac8806d7c0c6d3f8f"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.344285 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-xz8qb" Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.346841 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r5m97" event={"ID":"89658bb0-5353-4e67-95cd-cdb040086903","Type":"ContainerStarted","Data":"8d3fe41ca8fcbaad1000219b6b96a563f1a0d9586da77b4fc3c00749e4298700"} Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.383204 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2bmg9"] Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.395739 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2bmg9"] Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.413236 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xz8qb"] Oct 02 09:43:30 crc kubenswrapper[5035]: I1002 09:43:30.419687 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-xz8qb"] Oct 02 09:43:31 crc kubenswrapper[5035]: I1002 09:43:31.175351 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 09:43:31 crc kubenswrapper[5035]: W1002 09:43:31.654145 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0578ddf1_8297_4720_b0c1_8fa7b7dce11d.slice/crio-707a591b6c6c17c048cba134a31463c97ce5cc32938d9c39f79e658288392856 WatchSource:0}: Error finding container 707a591b6c6c17c048cba134a31463c97ce5cc32938d9c39f79e658288392856: Status 404 returned error can't find the container with id 707a591b6c6c17c048cba134a31463c97ce5cc32938d9c39f79e658288392856 Oct 02 09:43:32 crc kubenswrapper[5035]: I1002 09:43:32.172686 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2" path="/var/lib/kubelet/pods/68c2d2e7-4fa1-42d5-a964-1e9b7c0dbef2/volumes" Oct 02 09:43:32 crc kubenswrapper[5035]: I1002 09:43:32.173050 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8b9019d-0c99-4070-88d1-156f57617e20" path="/var/lib/kubelet/pods/c8b9019d-0c99-4070-88d1-156f57617e20/volumes" Oct 02 09:43:32 crc kubenswrapper[5035]: I1002 09:43:32.385924 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0578ddf1-8297-4720-b0c1-8fa7b7dce11d","Type":"ContainerStarted","Data":"707a591b6c6c17c048cba134a31463c97ce5cc32938d9c39f79e658288392856"} Oct 02 09:43:40 crc kubenswrapper[5035]: I1002 09:43:40.450188 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b08491fc-549c-4176-a8e8-5f733d838e3d","Type":"ContainerStarted","Data":"6fff0c76b83bf0d2ec43dbcb9be45b0161c660339ffebc6b68f3eb838a966ea4"} Oct 02 09:43:40 crc kubenswrapper[5035]: I1002 09:43:40.451768 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 09:43:40 crc kubenswrapper[5035]: I1002 09:43:40.454021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"da088dc0-8011-44e7-973b-66e75f33549c","Type":"ContainerStarted","Data":"e46eba542f32a1f802f1898b697cf62ba9397b6216b77f84febd1cb7e1f40112"} Oct 02 09:43:40 crc kubenswrapper[5035]: I1002 09:43:40.456589 5035 generic.go:334] "Generic (PLEG): container finished" podID="89658bb0-5353-4e67-95cd-cdb040086903" containerID="07c3ea471cf72ce31ad87329f254e26bd9dd9ac50f81dc48ad03030b8d9eacb2" exitCode=0 Oct 02 09:43:40 crc kubenswrapper[5035]: I1002 09:43:40.456733 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r5m97" event={"ID":"89658bb0-5353-4e67-95cd-cdb040086903","Type":"ContainerDied","Data":"07c3ea471cf72ce31ad87329f254e26bd9dd9ac50f81dc48ad03030b8d9eacb2"} Oct 02 09:43:40 crc kubenswrapper[5035]: I1002 09:43:40.460120 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9f58a85f-3d79-4ed1-a856-c491969c730f","Type":"ContainerStarted","Data":"073182ddcb39e8800e0f6a58c2c057897a2896e22dc33dfe358898fc31e5f9fc"} Oct 02 09:43:40 crc kubenswrapper[5035]: I1002 09:43:40.471810 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.870607477 podStartE2EDuration="24.47178901s" podCreationTimestamp="2025-10-02 09:43:16 +0000 UTC" firstStartedPulling="2025-10-02 09:43:29.458579793 +0000 UTC m=+974.814923818" lastFinishedPulling="2025-10-02 09:43:39.059761326 +0000 UTC m=+984.416105351" observedRunningTime="2025-10-02 09:43:40.471680076 +0000 UTC m=+985.828024111" watchObservedRunningTime="2025-10-02 09:43:40.47178901 +0000 UTC m=+985.828133035" Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.469767 5035 generic.go:334] "Generic (PLEG): container finished" podID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerID="6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8" exitCode=0 Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.469854 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" event={"ID":"c4138bce-f108-4a2c-b2fa-8a44f264ccc7","Type":"ContainerDied","Data":"6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8"} Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.474710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"07a982ff-c270-4133-a1a0-612dc7317d4b","Type":"ContainerStarted","Data":"e0c40baf5a464d5f352850d446e3fc8c968c1ebcd95fa8170cee2fdf1d3e2e55"} Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.481271 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r5m97" event={"ID":"89658bb0-5353-4e67-95cd-cdb040086903","Type":"ContainerStarted","Data":"8489e04e72f4504c3a4e693562b4bb5575be7aa121b29a95c1fb8ab19a810519"} Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.481318 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-r5m97" event={"ID":"89658bb0-5353-4e67-95cd-cdb040086903","Type":"ContainerStarted","Data":"e8fdaaac559d783b3f093973a5d7864ceaab969674b3b7c40ca110adf9b9614e"} Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.481825 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.482098 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.483441 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0578ddf1-8297-4720-b0c1-8fa7b7dce11d","Type":"ContainerStarted","Data":"b4cbe591ebc7288984a5ef005562b4ba181ecdb219153d05fb0ae158e3006e84"} Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.485541 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts" event={"ID":"57ea64fd-e931-45ab-9f0e-678c3c9ba7a2","Type":"ContainerStarted","Data":"3bbf7811c9704ec5a02ce8193e8063837489fc5796d7d74927fc288583b8606b"} Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.485699 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jvdts" Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.491024 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f677fa29-64c1-45f7-8f6d-c4c15f83db60","Type":"ContainerStarted","Data":"4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535"} Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.510745 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jvdts" podStartSLOduration=11.063363875 podStartE2EDuration="20.510720858s" podCreationTimestamp="2025-10-02 09:43:21 +0000 UTC" firstStartedPulling="2025-10-02 09:43:29.737007502 +0000 UTC m=+975.093351547" lastFinishedPulling="2025-10-02 09:43:39.184364505 +0000 UTC m=+984.540708530" observedRunningTime="2025-10-02 09:43:41.505501136 +0000 UTC m=+986.861845171" watchObservedRunningTime="2025-10-02 09:43:41.510720858 +0000 UTC m=+986.867064883" Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.530630 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-r5m97" podStartSLOduration=11.369918675 podStartE2EDuration="20.530612468s" podCreationTimestamp="2025-10-02 09:43:21 +0000 UTC" firstStartedPulling="2025-10-02 09:43:29.898363152 +0000 UTC m=+975.254707177" lastFinishedPulling="2025-10-02 09:43:39.059056955 +0000 UTC m=+984.415400970" observedRunningTime="2025-10-02 09:43:41.528388463 +0000 UTC m=+986.884732488" watchObservedRunningTime="2025-10-02 09:43:41.530612468 +0000 UTC m=+986.886956503" Oct 02 09:43:41 crc kubenswrapper[5035]: I1002 09:43:41.551780 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=12.346806213 podStartE2EDuration="23.551760474s" podCreationTimestamp="2025-10-02 09:43:18 +0000 UTC" firstStartedPulling="2025-10-02 09:43:29.543012942 +0000 UTC m=+974.899356967" lastFinishedPulling="2025-10-02 09:43:40.747967203 +0000 UTC m=+986.104311228" observedRunningTime="2025-10-02 09:43:41.543461182 +0000 UTC m=+986.899805207" watchObservedRunningTime="2025-10-02 09:43:41.551760474 +0000 UTC m=+986.908104499" Oct 02 09:43:42 crc kubenswrapper[5035]: I1002 09:43:42.503257 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3d7eafce-151f-43f3-8f74-cbfc9f368643","Type":"ContainerStarted","Data":"f1a518f45103e2a6dd09b8cc44fd94b3fc5bcbbd39842e6c0d7bc2c73063a1f3"} Oct 02 09:43:42 crc kubenswrapper[5035]: I1002 09:43:42.507304 5035 generic.go:334] "Generic (PLEG): container finished" podID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerID="713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf" exitCode=0 Oct 02 09:43:42 crc kubenswrapper[5035]: I1002 09:43:42.507346 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" event={"ID":"59c56020-0f32-4362-8ce8-a66ed3d219b4","Type":"ContainerDied","Data":"713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf"} Oct 02 09:43:42 crc kubenswrapper[5035]: I1002 09:43:42.512200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" event={"ID":"c4138bce-f108-4a2c-b2fa-8a44f264ccc7","Type":"ContainerStarted","Data":"5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475"} Oct 02 09:43:42 crc kubenswrapper[5035]: I1002 09:43:42.512543 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:42 crc kubenswrapper[5035]: I1002 09:43:42.515874 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 09:43:42 crc kubenswrapper[5035]: I1002 09:43:42.541802 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" podStartSLOduration=3.001859213 podStartE2EDuration="30.541785928s" podCreationTimestamp="2025-10-02 09:43:12 +0000 UTC" firstStartedPulling="2025-10-02 09:43:13.209988965 +0000 UTC m=+958.566332990" lastFinishedPulling="2025-10-02 09:43:40.74991568 +0000 UTC m=+986.106259705" observedRunningTime="2025-10-02 09:43:42.53740491 +0000 UTC m=+987.893748935" watchObservedRunningTime="2025-10-02 09:43:42.541785928 +0000 UTC m=+987.898129953" Oct 02 09:43:43 crc kubenswrapper[5035]: I1002 09:43:43.519506 5035 generic.go:334] "Generic (PLEG): container finished" podID="da088dc0-8011-44e7-973b-66e75f33549c" containerID="e46eba542f32a1f802f1898b697cf62ba9397b6216b77f84febd1cb7e1f40112" exitCode=0 Oct 02 09:43:43 crc kubenswrapper[5035]: I1002 09:43:43.519564 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"da088dc0-8011-44e7-973b-66e75f33549c","Type":"ContainerDied","Data":"e46eba542f32a1f802f1898b697cf62ba9397b6216b77f84febd1cb7e1f40112"} Oct 02 09:43:43 crc kubenswrapper[5035]: I1002 09:43:43.522273 5035 generic.go:334] "Generic (PLEG): container finished" podID="9f58a85f-3d79-4ed1-a856-c491969c730f" containerID="073182ddcb39e8800e0f6a58c2c057897a2896e22dc33dfe358898fc31e5f9fc" exitCode=0 Oct 02 09:43:43 crc kubenswrapper[5035]: I1002 09:43:43.522599 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9f58a85f-3d79-4ed1-a856-c491969c730f","Type":"ContainerDied","Data":"073182ddcb39e8800e0f6a58c2c057897a2896e22dc33dfe358898fc31e5f9fc"} Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.532137 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" event={"ID":"59c56020-0f32-4362-8ce8-a66ed3d219b4","Type":"ContainerStarted","Data":"ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873"} Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.532581 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.534463 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"07a982ff-c270-4133-a1a0-612dc7317d4b","Type":"ContainerStarted","Data":"b10dd01ee5f9bf7bdd152543175e79a9578ea40c22e9f93130ad1b6fb7249ccf"} Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.536942 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"da088dc0-8011-44e7-973b-66e75f33549c","Type":"ContainerStarted","Data":"84c54587fe15d73778f5a1fd77cd5e00d5f2971cd197fa7a1bf3a0566fa657fd"} Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.539287 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"0578ddf1-8297-4720-b0c1-8fa7b7dce11d","Type":"ContainerStarted","Data":"d1e4d9aad984ac969b114e587f52192e1ab521d69ebe9daff64df58f6648abc7"} Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.541834 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"9f58a85f-3d79-4ed1-a856-c491969c730f","Type":"ContainerStarted","Data":"6a0ff0bf842c8d6a080ff324c64be6de16665db5ec4c54ec6d75682bdf006296"} Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.575361 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" podStartSLOduration=-9223372004.279436 podStartE2EDuration="32.575340054s" podCreationTimestamp="2025-10-02 09:43:12 +0000 UTC" firstStartedPulling="2025-10-02 09:43:13.451055494 +0000 UTC m=+958.807399519" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:43:44.552275723 +0000 UTC m=+989.908619768" watchObservedRunningTime="2025-10-02 09:43:44.575340054 +0000 UTC m=+989.931684139" Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.585323 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=9.923183979000001 podStartE2EDuration="23.585303944s" podCreationTimestamp="2025-10-02 09:43:21 +0000 UTC" firstStartedPulling="2025-10-02 09:43:30.132683826 +0000 UTC m=+975.489027851" lastFinishedPulling="2025-10-02 09:43:43.794803791 +0000 UTC m=+989.151147816" observedRunningTime="2025-10-02 09:43:44.579183766 +0000 UTC m=+989.935527811" watchObservedRunningTime="2025-10-02 09:43:44.585303944 +0000 UTC m=+989.941647969" Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.609638 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.166290428 podStartE2EDuration="29.609469288s" podCreationTimestamp="2025-10-02 09:43:15 +0000 UTC" firstStartedPulling="2025-10-02 09:43:29.739183346 +0000 UTC m=+975.095527371" lastFinishedPulling="2025-10-02 09:43:39.182362206 +0000 UTC m=+984.538706231" observedRunningTime="2025-10-02 09:43:44.606250765 +0000 UTC m=+989.962594840" watchObservedRunningTime="2025-10-02 09:43:44.609469288 +0000 UTC m=+989.965813313" Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.625019 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.477984283 podStartE2EDuration="19.62499383s" podCreationTimestamp="2025-10-02 09:43:25 +0000 UTC" firstStartedPulling="2025-10-02 09:43:31.658746273 +0000 UTC m=+977.015090298" lastFinishedPulling="2025-10-02 09:43:43.80575582 +0000 UTC m=+989.162099845" observedRunningTime="2025-10-02 09:43:44.620430448 +0000 UTC m=+989.976774483" watchObservedRunningTime="2025-10-02 09:43:44.62499383 +0000 UTC m=+989.981337865" Oct 02 09:43:44 crc kubenswrapper[5035]: I1002 09:43:44.649633 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.8741522 podStartE2EDuration="30.649588877s" podCreationTimestamp="2025-10-02 09:43:14 +0000 UTC" firstStartedPulling="2025-10-02 09:43:29.315727303 +0000 UTC m=+974.672071328" lastFinishedPulling="2025-10-02 09:43:39.09116397 +0000 UTC m=+984.447508005" observedRunningTime="2025-10-02 09:43:44.644251421 +0000 UTC m=+990.000595446" watchObservedRunningTime="2025-10-02 09:43:44.649588877 +0000 UTC m=+990.005932932" Oct 02 09:43:45 crc kubenswrapper[5035]: I1002 09:43:45.643098 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 09:43:45 crc kubenswrapper[5035]: I1002 09:43:45.643423 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 09:43:46 crc kubenswrapper[5035]: I1002 09:43:46.522751 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:46 crc kubenswrapper[5035]: I1002 09:43:46.660272 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:46 crc kubenswrapper[5035]: I1002 09:43:46.706523 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.000759 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.000839 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.317241 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.523400 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.569141 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.586072 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.625250 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.661759 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.713879 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.817311 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.897369 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9bs5d"] Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.897722 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerName="dnsmasq-dns" containerID="cri-o://ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873" gracePeriod=10 Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.930626 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2mkm7"] Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.932127 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.936337 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 09:43:47 crc kubenswrapper[5035]: I1002 09:43:47.946704 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2mkm7"] Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.015505 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-v56nx"] Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.016616 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.037020 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.069468 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.069560 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-config\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.069588 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.069789 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg5f6\" (UniqueName: \"kubernetes.io/projected/934387a8-9eaa-4044-a1aa-17bdcdf89902-kube-api-access-sg5f6\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.103602 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v56nx"] Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175346 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg5f6\" (UniqueName: \"kubernetes.io/projected/934387a8-9eaa-4044-a1aa-17bdcdf89902-kube-api-access-sg5f6\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175438 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92aa4b07-2174-4435-b84c-426260df54fb-combined-ca-bundle\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175477 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175500 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92aa4b07-2174-4435-b84c-426260df54fb-config\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175548 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-config\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175566 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175606 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/92aa4b07-2174-4435-b84c-426260df54fb-ovs-rundir\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175628 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/92aa4b07-2174-4435-b84c-426260df54fb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175648 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/92aa4b07-2174-4435-b84c-426260df54fb-ovn-rundir\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.175664 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mjd7\" (UniqueName: \"kubernetes.io/projected/92aa4b07-2174-4435-b84c-426260df54fb-kube-api-access-7mjd7\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.176855 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.177000 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.177379 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-config\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.220371 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg5f6\" (UniqueName: \"kubernetes.io/projected/934387a8-9eaa-4044-a1aa-17bdcdf89902-kube-api-access-sg5f6\") pod \"dnsmasq-dns-7fd796d7df-2mkm7\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.253374 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.285066 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92aa4b07-2174-4435-b84c-426260df54fb-combined-ca-bundle\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.285465 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92aa4b07-2174-4435-b84c-426260df54fb-config\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.285631 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/92aa4b07-2174-4435-b84c-426260df54fb-ovs-rundir\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.285663 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/92aa4b07-2174-4435-b84c-426260df54fb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.285692 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mjd7\" (UniqueName: \"kubernetes.io/projected/92aa4b07-2174-4435-b84c-426260df54fb-kube-api-access-7mjd7\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.285716 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/92aa4b07-2174-4435-b84c-426260df54fb-ovn-rundir\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.287281 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/92aa4b07-2174-4435-b84c-426260df54fb-ovs-rundir\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.288275 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92aa4b07-2174-4435-b84c-426260df54fb-config\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.289037 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/92aa4b07-2174-4435-b84c-426260df54fb-ovn-rundir\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.290692 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92aa4b07-2174-4435-b84c-426260df54fb-combined-ca-bundle\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.294827 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/92aa4b07-2174-4435-b84c-426260df54fb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.315163 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mjd7\" (UniqueName: \"kubernetes.io/projected/92aa4b07-2174-4435-b84c-426260df54fb-kube-api-access-7mjd7\") pod \"ovn-controller-metrics-v56nx\" (UID: \"92aa4b07-2174-4435-b84c-426260df54fb\") " pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.366685 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v56nx" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.412916 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2mkm7"] Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.478877 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.481655 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9swnc"] Oct 02 09:43:48 crc kubenswrapper[5035]: E1002 09:43:48.482251 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerName="init" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.482353 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerName="init" Oct 02 09:43:48 crc kubenswrapper[5035]: E1002 09:43:48.506255 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerName="dnsmasq-dns" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.506634 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerName="dnsmasq-dns" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.507004 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerName="dnsmasq-dns" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.513367 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.526511 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.544433 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9swnc"] Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.592131 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-dns-svc\") pod \"59c56020-0f32-4362-8ce8-a66ed3d219b4\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.592209 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-config\") pod \"59c56020-0f32-4362-8ce8-a66ed3d219b4\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.592263 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7krhs\" (UniqueName: \"kubernetes.io/projected/59c56020-0f32-4362-8ce8-a66ed3d219b4-kube-api-access-7krhs\") pod \"59c56020-0f32-4362-8ce8-a66ed3d219b4\" (UID: \"59c56020-0f32-4362-8ce8-a66ed3d219b4\") " Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.602729 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59c56020-0f32-4362-8ce8-a66ed3d219b4-kube-api-access-7krhs" (OuterVolumeSpecName: "kube-api-access-7krhs") pod "59c56020-0f32-4362-8ce8-a66ed3d219b4" (UID: "59c56020-0f32-4362-8ce8-a66ed3d219b4"). InnerVolumeSpecName "kube-api-access-7krhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.610713 5035 generic.go:334] "Generic (PLEG): container finished" podID="59c56020-0f32-4362-8ce8-a66ed3d219b4" containerID="ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873" exitCode=0 Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.610821 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.610873 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" event={"ID":"59c56020-0f32-4362-8ce8-a66ed3d219b4","Type":"ContainerDied","Data":"ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873"} Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.610916 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-9bs5d" event={"ID":"59c56020-0f32-4362-8ce8-a66ed3d219b4","Type":"ContainerDied","Data":"bce56cfed1b215af4b1f4b1d5a6eb601b6f75651ab0dee0ad01cda423e484008"} Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.610939 5035 scope.go:117] "RemoveContainer" containerID="ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.657639 5035 scope.go:117] "RemoveContainer" containerID="713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.658732 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59c56020-0f32-4362-8ce8-a66ed3d219b4" (UID: "59c56020-0f32-4362-8ce8-a66ed3d219b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.662993 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-config" (OuterVolumeSpecName: "config") pod "59c56020-0f32-4362-8ce8-a66ed3d219b4" (UID: "59c56020-0f32-4362-8ce8-a66ed3d219b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.698347 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699111 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjljh\" (UniqueName: \"kubernetes.io/projected/8db39c9d-ae09-428d-86af-993f983e8681-kube-api-access-pjljh\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699165 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699225 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699283 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-config\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699316 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699369 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699384 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c56020-0f32-4362-8ce8-a66ed3d219b4-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.699397 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7krhs\" (UniqueName: \"kubernetes.io/projected/59c56020-0f32-4362-8ce8-a66ed3d219b4-kube-api-access-7krhs\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.708989 5035 scope.go:117] "RemoveContainer" containerID="ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873" Oct 02 09:43:48 crc kubenswrapper[5035]: E1002 09:43:48.710497 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873\": container with ID starting with ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873 not found: ID does not exist" containerID="ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.710699 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873"} err="failed to get container status \"ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873\": rpc error: code = NotFound desc = could not find container \"ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873\": container with ID starting with ae985a9fdc6f59caa7a5b9048c572b4eaa8cb89af12f6d350fac9d1ba522a873 not found: ID does not exist" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.710807 5035 scope.go:117] "RemoveContainer" containerID="713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf" Oct 02 09:43:48 crc kubenswrapper[5035]: E1002 09:43:48.711130 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf\": container with ID starting with 713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf not found: ID does not exist" containerID="713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.711168 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf"} err="failed to get container status \"713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf\": rpc error: code = NotFound desc = could not find container \"713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf\": container with ID starting with 713708aa3130156c4cf98526f7d552730e0ca312ed55f1fdd3bf688c0cf747cf not found: ID does not exist" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.800379 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.800874 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-config\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.800900 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.801112 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjljh\" (UniqueName: \"kubernetes.io/projected/8db39c9d-ae09-428d-86af-993f983e8681-kube-api-access-pjljh\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.801165 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.801977 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.802738 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-config\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.802784 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.803422 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.837148 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjljh\" (UniqueName: \"kubernetes.io/projected/8db39c9d-ae09-428d-86af-993f983e8681-kube-api-access-pjljh\") pod \"dnsmasq-dns-86db49b7ff-9swnc\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.892911 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.893896 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.924907 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.929606 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.929744 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.929851 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.934258 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-f5h6m" Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.947157 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2mkm7"] Oct 02 09:43:48 crc kubenswrapper[5035]: I1002 09:43:48.964845 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.006696 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9bs5d"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.012743 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-9bs5d"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.013518 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.013584 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fe804ba-727c-49dd-a12b-43ff832db070-config\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.013630 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3fe804ba-727c-49dd-a12b-43ff832db070-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.017200 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.017282 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fe804ba-727c-49dd-a12b-43ff832db070-scripts\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.017570 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.017610 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwktg\" (UniqueName: \"kubernetes.io/projected/3fe804ba-727c-49dd-a12b-43ff832db070-kube-api-access-rwktg\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.057170 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v56nx"] Oct 02 09:43:49 crc kubenswrapper[5035]: W1002 09:43:49.075216 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92aa4b07_2174_4435_b84c_426260df54fb.slice/crio-29fb57db4f0d1e2b2e098d6c1ce880b83883e75e7827fcf84215eee9244c1421 WatchSource:0}: Error finding container 29fb57db4f0d1e2b2e098d6c1ce880b83883e75e7827fcf84215eee9244c1421: Status 404 returned error can't find the container with id 29fb57db4f0d1e2b2e098d6c1ce880b83883e75e7827fcf84215eee9244c1421 Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.099191 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.118718 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.118787 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fe804ba-727c-49dd-a12b-43ff832db070-scripts\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.118826 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.118851 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwktg\" (UniqueName: \"kubernetes.io/projected/3fe804ba-727c-49dd-a12b-43ff832db070-kube-api-access-rwktg\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.118902 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.118925 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fe804ba-727c-49dd-a12b-43ff832db070-config\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.118968 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3fe804ba-727c-49dd-a12b-43ff832db070-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.119660 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3fe804ba-727c-49dd-a12b-43ff832db070-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.119955 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3fe804ba-727c-49dd-a12b-43ff832db070-scripts\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.121900 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3fe804ba-727c-49dd-a12b-43ff832db070-config\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.122599 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.124481 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.126112 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3fe804ba-727c-49dd-a12b-43ff832db070-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.151066 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwktg\" (UniqueName: \"kubernetes.io/projected/3fe804ba-727c-49dd-a12b-43ff832db070-kube-api-access-rwktg\") pod \"ovn-northd-0\" (UID: \"3fe804ba-727c-49dd-a12b-43ff832db070\") " pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.170068 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.283699 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.299936 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.435446 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9swnc"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.479963 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-xcb6x"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.481265 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.496662 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-xcb6x"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.518382 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9swnc"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.621965 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v56nx" event={"ID":"92aa4b07-2174-4435-b84c-426260df54fb","Type":"ContainerStarted","Data":"cbaf1aada98e41641dbe770c9408932b86fbbea4e8ec196dba946a0f639fd8ac"} Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.622078 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v56nx" event={"ID":"92aa4b07-2174-4435-b84c-426260df54fb","Type":"ContainerStarted","Data":"29fb57db4f0d1e2b2e098d6c1ce880b83883e75e7827fcf84215eee9244c1421"} Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.627415 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" event={"ID":"8db39c9d-ae09-428d-86af-993f983e8681","Type":"ContainerStarted","Data":"aaf5f829a9fbea8eb50ebed9e2eca02266127687d8fd58f4a594407471fdc3e6"} Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.629487 5035 generic.go:334] "Generic (PLEG): container finished" podID="934387a8-9eaa-4044-a1aa-17bdcdf89902" containerID="4319be6d97f7574feb02160b2358d743be5091290f614655ce3628658df944cf" exitCode=0 Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.630807 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" event={"ID":"934387a8-9eaa-4044-a1aa-17bdcdf89902","Type":"ContainerDied","Data":"4319be6d97f7574feb02160b2358d743be5091290f614655ce3628658df944cf"} Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.630865 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" event={"ID":"934387a8-9eaa-4044-a1aa-17bdcdf89902","Type":"ContainerStarted","Data":"ead71a73318d4d05f1ac3cc71e991d2a05e57e4906f45bce9949d54d5968465f"} Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.638634 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-dns-svc\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.638724 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.638842 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-config\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.638988 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.639041 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fstp9\" (UniqueName: \"kubernetes.io/projected/30d3793a-815a-4e7b-b3ec-6300de91a736-kube-api-access-fstp9\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.651430 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-v56nx" podStartSLOduration=2.651398093 podStartE2EDuration="2.651398093s" podCreationTimestamp="2025-10-02 09:43:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:43:49.640628009 +0000 UTC m=+994.996972034" watchObservedRunningTime="2025-10-02 09:43:49.651398093 +0000 UTC m=+995.007742118" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.741154 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-dns-svc\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.741371 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.741435 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-config\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.741556 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.741625 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fstp9\" (UniqueName: \"kubernetes.io/projected/30d3793a-815a-4e7b-b3ec-6300de91a736-kube-api-access-fstp9\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.743677 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-dns-svc\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.745101 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-config\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.746463 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.747085 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.755213 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.770173 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fstp9\" (UniqueName: \"kubernetes.io/projected/30d3793a-815a-4e7b-b3ec-6300de91a736-kube-api-access-fstp9\") pod \"dnsmasq-dns-698758b865-xcb6x\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.884029 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:49 crc kubenswrapper[5035]: I1002 09:43:49.982914 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.049135 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-config\") pod \"934387a8-9eaa-4044-a1aa-17bdcdf89902\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.049226 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg5f6\" (UniqueName: \"kubernetes.io/projected/934387a8-9eaa-4044-a1aa-17bdcdf89902-kube-api-access-sg5f6\") pod \"934387a8-9eaa-4044-a1aa-17bdcdf89902\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.049339 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-ovsdbserver-nb\") pod \"934387a8-9eaa-4044-a1aa-17bdcdf89902\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.049378 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-dns-svc\") pod \"934387a8-9eaa-4044-a1aa-17bdcdf89902\" (UID: \"934387a8-9eaa-4044-a1aa-17bdcdf89902\") " Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.072883 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/934387a8-9eaa-4044-a1aa-17bdcdf89902-kube-api-access-sg5f6" (OuterVolumeSpecName: "kube-api-access-sg5f6") pod "934387a8-9eaa-4044-a1aa-17bdcdf89902" (UID: "934387a8-9eaa-4044-a1aa-17bdcdf89902"). InnerVolumeSpecName "kube-api-access-sg5f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.135083 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "934387a8-9eaa-4044-a1aa-17bdcdf89902" (UID: "934387a8-9eaa-4044-a1aa-17bdcdf89902"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.147139 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-config" (OuterVolumeSpecName: "config") pod "934387a8-9eaa-4044-a1aa-17bdcdf89902" (UID: "934387a8-9eaa-4044-a1aa-17bdcdf89902"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.162759 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.162789 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.162801 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg5f6\" (UniqueName: \"kubernetes.io/projected/934387a8-9eaa-4044-a1aa-17bdcdf89902-kube-api-access-sg5f6\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.171134 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "934387a8-9eaa-4044-a1aa-17bdcdf89902" (UID: "934387a8-9eaa-4044-a1aa-17bdcdf89902"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.196464 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59c56020-0f32-4362-8ce8-a66ed3d219b4" path="/var/lib/kubelet/pods/59c56020-0f32-4362-8ce8-a66ed3d219b4/volumes" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.282522 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/934387a8-9eaa-4044-a1aa-17bdcdf89902-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.480780 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-xcb6x"] Oct 02 09:43:50 crc kubenswrapper[5035]: W1002 09:43:50.485028 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30d3793a_815a_4e7b_b3ec_6300de91a736.slice/crio-6d03e94f44cd4114a77fe1b39792fe7c541f259ffa07fbffa739ac0c8dd1a7f2 WatchSource:0}: Error finding container 6d03e94f44cd4114a77fe1b39792fe7c541f259ffa07fbffa739ac0c8dd1a7f2: Status 404 returned error can't find the container with id 6d03e94f44cd4114a77fe1b39792fe7c541f259ffa07fbffa739ac0c8dd1a7f2 Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.612937 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 09:43:50 crc kubenswrapper[5035]: E1002 09:43:50.613519 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="934387a8-9eaa-4044-a1aa-17bdcdf89902" containerName="init" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.613586 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="934387a8-9eaa-4044-a1aa-17bdcdf89902" containerName="init" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.613812 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="934387a8-9eaa-4044-a1aa-17bdcdf89902" containerName="init" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.621302 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.625410 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.625693 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.626139 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-fzkr5" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.626348 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.628576 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.655668 5035 generic.go:334] "Generic (PLEG): container finished" podID="8db39c9d-ae09-428d-86af-993f983e8681" containerID="b2ed4d10cc0bb105525cda0c2e16a4752fb5966e770e24ddc8b4a007ccb55e73" exitCode=0 Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.655985 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" event={"ID":"8db39c9d-ae09-428d-86af-993f983e8681","Type":"ContainerDied","Data":"b2ed4d10cc0bb105525cda0c2e16a4752fb5966e770e24ddc8b4a007ccb55e73"} Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.658469 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" event={"ID":"934387a8-9eaa-4044-a1aa-17bdcdf89902","Type":"ContainerDied","Data":"ead71a73318d4d05f1ac3cc71e991d2a05e57e4906f45bce9949d54d5968465f"} Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.658657 5035 scope.go:117] "RemoveContainer" containerID="4319be6d97f7574feb02160b2358d743be5091290f614655ce3628658df944cf" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.659285 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-2mkm7" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.661494 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-xcb6x" event={"ID":"30d3793a-815a-4e7b-b3ec-6300de91a736","Type":"ContainerStarted","Data":"6d03e94f44cd4114a77fe1b39792fe7c541f259ffa07fbffa739ac0c8dd1a7f2"} Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.671005 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3fe804ba-727c-49dd-a12b-43ff832db070","Type":"ContainerStarted","Data":"72867a871f10a906f9f18fc0d67139f49041b62796e380b59869791f6cf1dab6"} Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.778447 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2mkm7"] Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.788462 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-2mkm7"] Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.795348 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mwnh\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-kube-api-access-8mwnh\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.795389 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2aaa2a11-2896-4690-819b-61c6069ea6db-cache\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.795490 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.795551 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2aaa2a11-2896-4690-819b-61c6069ea6db-lock\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.795567 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.896484 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mwnh\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-kube-api-access-8mwnh\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.896559 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2aaa2a11-2896-4690-819b-61c6069ea6db-cache\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.896682 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.896724 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2aaa2a11-2896-4690-819b-61c6069ea6db-lock\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.896745 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: E1002 09:43:50.896925 5035 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:43:50 crc kubenswrapper[5035]: E1002 09:43:50.896942 5035 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:43:50 crc kubenswrapper[5035]: E1002 09:43:50.896993 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift podName:2aaa2a11-2896-4690-819b-61c6069ea6db nodeName:}" failed. No retries permitted until 2025-10-02 09:43:51.396972369 +0000 UTC m=+996.753316394 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift") pod "swift-storage-0" (UID: "2aaa2a11-2896-4690-819b-61c6069ea6db") : configmap "swift-ring-files" not found Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.897950 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.898165 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/2aaa2a11-2896-4690-819b-61c6069ea6db-lock\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.898603 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/2aaa2a11-2896-4690-819b-61c6069ea6db-cache\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.917624 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mwnh\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-kube-api-access-8mwnh\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.923906 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:50 crc kubenswrapper[5035]: I1002 09:43:50.974151 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.099908 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-nb\") pod \"8db39c9d-ae09-428d-86af-993f983e8681\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.100002 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-sb\") pod \"8db39c9d-ae09-428d-86af-993f983e8681\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.100152 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-dns-svc\") pod \"8db39c9d-ae09-428d-86af-993f983e8681\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.100227 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjljh\" (UniqueName: \"kubernetes.io/projected/8db39c9d-ae09-428d-86af-993f983e8681-kube-api-access-pjljh\") pod \"8db39c9d-ae09-428d-86af-993f983e8681\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.100285 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-config\") pod \"8db39c9d-ae09-428d-86af-993f983e8681\" (UID: \"8db39c9d-ae09-428d-86af-993f983e8681\") " Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.118131 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wcbql"] Oct 02 09:43:51 crc kubenswrapper[5035]: E1002 09:43:51.118571 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8db39c9d-ae09-428d-86af-993f983e8681" containerName="init" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.118588 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8db39c9d-ae09-428d-86af-993f983e8681" containerName="init" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.118812 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8db39c9d-ae09-428d-86af-993f983e8681" containerName="init" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.119450 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.123524 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.123623 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.123911 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.151281 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8db39c9d-ae09-428d-86af-993f983e8681-kube-api-access-pjljh" (OuterVolumeSpecName: "kube-api-access-pjljh") pod "8db39c9d-ae09-428d-86af-993f983e8681" (UID: "8db39c9d-ae09-428d-86af-993f983e8681"). InnerVolumeSpecName "kube-api-access-pjljh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.151907 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8db39c9d-ae09-428d-86af-993f983e8681" (UID: "8db39c9d-ae09-428d-86af-993f983e8681"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.151914 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8db39c9d-ae09-428d-86af-993f983e8681" (UID: "8db39c9d-ae09-428d-86af-993f983e8681"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.154068 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-config" (OuterVolumeSpecName: "config") pod "8db39c9d-ae09-428d-86af-993f983e8681" (UID: "8db39c9d-ae09-428d-86af-993f983e8681"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.157082 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8db39c9d-ae09-428d-86af-993f983e8681" (UID: "8db39c9d-ae09-428d-86af-993f983e8681"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.169923 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wcbql"] Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.202593 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.202634 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjljh\" (UniqueName: \"kubernetes.io/projected/8db39c9d-ae09-428d-86af-993f983e8681-kube-api-access-pjljh\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.202678 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.202699 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.202711 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8db39c9d-ae09-428d-86af-993f983e8681-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.303776 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-swiftconf\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.303851 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-etc-swift\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.303882 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-scripts\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.303904 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sksfz\" (UniqueName: \"kubernetes.io/projected/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-kube-api-access-sksfz\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.304023 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-ring-data-devices\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.304078 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-combined-ca-bundle\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.304128 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-dispersionconf\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.405787 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-swiftconf\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.408961 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-etc-swift\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.409004 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-scripts\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.409654 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sksfz\" (UniqueName: \"kubernetes.io/projected/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-kube-api-access-sksfz\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.409563 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-etc-swift\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.409900 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-scripts\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.410213 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-ring-data-devices\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.410293 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-combined-ca-bundle\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.416299 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-swiftconf\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.416375 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-combined-ca-bundle\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.424419 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-dispersionconf\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.424590 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:51 crc kubenswrapper[5035]: E1002 09:43:51.424864 5035 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:43:51 crc kubenswrapper[5035]: E1002 09:43:51.424919 5035 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:43:51 crc kubenswrapper[5035]: E1002 09:43:51.424984 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift podName:2aaa2a11-2896-4690-819b-61c6069ea6db nodeName:}" failed. No retries permitted until 2025-10-02 09:43:52.424962276 +0000 UTC m=+997.781306301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift") pod "swift-storage-0" (UID: "2aaa2a11-2896-4690-819b-61c6069ea6db") : configmap "swift-ring-files" not found Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.425195 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-ring-data-devices\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.428523 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-dispersionconf\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.433277 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sksfz\" (UniqueName: \"kubernetes.io/projected/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-kube-api-access-sksfz\") pod \"swift-ring-rebalance-wcbql\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.522560 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.682141 5035 generic.go:334] "Generic (PLEG): container finished" podID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerID="401a1e90b66d3b939e272acee7a4a073be94d571167a055f65910165f45b4664" exitCode=0 Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.682217 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-xcb6x" event={"ID":"30d3793a-815a-4e7b-b3ec-6300de91a736","Type":"ContainerDied","Data":"401a1e90b66d3b939e272acee7a4a073be94d571167a055f65910165f45b4664"} Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.687998 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" event={"ID":"8db39c9d-ae09-428d-86af-993f983e8681","Type":"ContainerDied","Data":"aaf5f829a9fbea8eb50ebed9e2eca02266127687d8fd58f4a594407471fdc3e6"} Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.688027 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-9swnc" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.688059 5035 scope.go:117] "RemoveContainer" containerID="b2ed4d10cc0bb105525cda0c2e16a4752fb5966e770e24ddc8b4a007ccb55e73" Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.777825 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9swnc"] Oct 02 09:43:51 crc kubenswrapper[5035]: I1002 09:43:51.793781 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-9swnc"] Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.000852 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wcbql"] Oct 02 09:43:52 crc kubenswrapper[5035]: W1002 09:43:52.001583 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65e1909b_5dc0_45ad_af61_f3cd8ce48ace.slice/crio-997822288cf8a9f6ec7535d1aa60c200ef29e0d536d366e90bdb392ff907fa37 WatchSource:0}: Error finding container 997822288cf8a9f6ec7535d1aa60c200ef29e0d536d366e90bdb392ff907fa37: Status 404 returned error can't find the container with id 997822288cf8a9f6ec7535d1aa60c200ef29e0d536d366e90bdb392ff907fa37 Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.176421 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8db39c9d-ae09-428d-86af-993f983e8681" path="/var/lib/kubelet/pods/8db39c9d-ae09-428d-86af-993f983e8681/volumes" Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.180975 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="934387a8-9eaa-4044-a1aa-17bdcdf89902" path="/var/lib/kubelet/pods/934387a8-9eaa-4044-a1aa-17bdcdf89902/volumes" Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.447860 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:52 crc kubenswrapper[5035]: E1002 09:43:52.448088 5035 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:43:52 crc kubenswrapper[5035]: E1002 09:43:52.448121 5035 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:43:52 crc kubenswrapper[5035]: E1002 09:43:52.448188 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift podName:2aaa2a11-2896-4690-819b-61c6069ea6db nodeName:}" failed. No retries permitted until 2025-10-02 09:43:54.448167357 +0000 UTC m=+999.804511382 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift") pod "swift-storage-0" (UID: "2aaa2a11-2896-4690-819b-61c6069ea6db") : configmap "swift-ring-files" not found Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.706923 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3fe804ba-727c-49dd-a12b-43ff832db070","Type":"ContainerStarted","Data":"e88b9912849762feed384b143ae6d14d65d5f712907b8b46941a24c9ee9900ac"} Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.706968 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3fe804ba-727c-49dd-a12b-43ff832db070","Type":"ContainerStarted","Data":"a34caf9368780ceb460e0a2e56f455f7b228cfb4a6bbc1638cab6f8848dad1fe"} Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.707028 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.714124 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wcbql" event={"ID":"65e1909b-5dc0-45ad-af61-f3cd8ce48ace","Type":"ContainerStarted","Data":"997822288cf8a9f6ec7535d1aa60c200ef29e0d536d366e90bdb392ff907fa37"} Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.717076 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-xcb6x" event={"ID":"30d3793a-815a-4e7b-b3ec-6300de91a736","Type":"ContainerStarted","Data":"ac79e8484380468920f1feaed79f4abc5c149f7365a87f83e5ef4a6dea908b60"} Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.717424 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.729675 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.710033265 podStartE2EDuration="4.729652875s" podCreationTimestamp="2025-10-02 09:43:48 +0000 UTC" firstStartedPulling="2025-10-02 09:43:49.747630155 +0000 UTC m=+995.103974180" lastFinishedPulling="2025-10-02 09:43:51.767249765 +0000 UTC m=+997.123593790" observedRunningTime="2025-10-02 09:43:52.725784813 +0000 UTC m=+998.082128858" watchObservedRunningTime="2025-10-02 09:43:52.729652875 +0000 UTC m=+998.085996900" Oct 02 09:43:52 crc kubenswrapper[5035]: I1002 09:43:52.749663 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-xcb6x" podStartSLOduration=3.7496456780000003 podStartE2EDuration="3.749645678s" podCreationTimestamp="2025-10-02 09:43:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:43:52.745704023 +0000 UTC m=+998.102048048" watchObservedRunningTime="2025-10-02 09:43:52.749645678 +0000 UTC m=+998.105989703" Oct 02 09:43:54 crc kubenswrapper[5035]: I1002 09:43:54.481440 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:54 crc kubenswrapper[5035]: E1002 09:43:54.481644 5035 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:43:54 crc kubenswrapper[5035]: E1002 09:43:54.481881 5035 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:43:54 crc kubenswrapper[5035]: E1002 09:43:54.481929 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift podName:2aaa2a11-2896-4690-819b-61c6069ea6db nodeName:}" failed. No retries permitted until 2025-10-02 09:43:58.481916829 +0000 UTC m=+1003.838260854 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift") pod "swift-storage-0" (UID: "2aaa2a11-2896-4690-819b-61c6069ea6db") : configmap "swift-ring-files" not found Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.538262 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.538335 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.538391 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.539236 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a1bca98fb1c6a4355a4320624f5591fb78dba746011bbeaf90858ccf387d2ab3"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.539311 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://a1bca98fb1c6a4355a4320624f5591fb78dba746011bbeaf90858ccf387d2ab3" gracePeriod=600 Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.743218 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="a1bca98fb1c6a4355a4320624f5591fb78dba746011bbeaf90858ccf387d2ab3" exitCode=0 Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.743260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"a1bca98fb1c6a4355a4320624f5591fb78dba746011bbeaf90858ccf387d2ab3"} Oct 02 09:43:55 crc kubenswrapper[5035]: I1002 09:43:55.743641 5035 scope.go:117] "RemoveContainer" containerID="f4d0414702a1d48c4861b01b90ad79bd9fac74730456a0cf8ab31e26b08e9e73" Oct 02 09:43:56 crc kubenswrapper[5035]: I1002 09:43:56.764301 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wcbql" event={"ID":"65e1909b-5dc0-45ad-af61-f3cd8ce48ace","Type":"ContainerStarted","Data":"b9967648c3c14b81f9899758a17798f6e0ac62004727202f37d54af50f16b134"} Oct 02 09:43:56 crc kubenswrapper[5035]: I1002 09:43:56.769257 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"db9ffa0950502b83608fcd73436cfa37aeaa8a721d02dc03e8e35488a5442a33"} Oct 02 09:43:56 crc kubenswrapper[5035]: I1002 09:43:56.788817 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-wcbql" podStartSLOduration=1.7052921749999999 podStartE2EDuration="5.788800956s" podCreationTimestamp="2025-10-02 09:43:51 +0000 UTC" firstStartedPulling="2025-10-02 09:43:52.003360442 +0000 UTC m=+997.359704457" lastFinishedPulling="2025-10-02 09:43:56.086869203 +0000 UTC m=+1001.443213238" observedRunningTime="2025-10-02 09:43:56.782438931 +0000 UTC m=+1002.138782976" watchObservedRunningTime="2025-10-02 09:43:56.788800956 +0000 UTC m=+1002.145144981" Oct 02 09:43:56 crc kubenswrapper[5035]: I1002 09:43:56.969751 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bqd4v"] Oct 02 09:43:56 crc kubenswrapper[5035]: I1002 09:43:56.973051 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bqd4v" Oct 02 09:43:56 crc kubenswrapper[5035]: I1002 09:43:56.981374 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bqd4v"] Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.122375 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s77kn\" (UniqueName: \"kubernetes.io/projected/cb09fa9f-7359-46a9-baf8-758474e12535-kube-api-access-s77kn\") pod \"keystone-db-create-bqd4v\" (UID: \"cb09fa9f-7359-46a9-baf8-758474e12535\") " pod="openstack/keystone-db-create-bqd4v" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.163857 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wc7vd"] Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.165012 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wc7vd" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.170917 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wc7vd"] Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.224104 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76fc7\" (UniqueName: \"kubernetes.io/projected/e56670c7-7dcd-4294-b852-cf1397b892f7-kube-api-access-76fc7\") pod \"placement-db-create-wc7vd\" (UID: \"e56670c7-7dcd-4294-b852-cf1397b892f7\") " pod="openstack/placement-db-create-wc7vd" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.224471 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s77kn\" (UniqueName: \"kubernetes.io/projected/cb09fa9f-7359-46a9-baf8-758474e12535-kube-api-access-s77kn\") pod \"keystone-db-create-bqd4v\" (UID: \"cb09fa9f-7359-46a9-baf8-758474e12535\") " pod="openstack/keystone-db-create-bqd4v" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.241340 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s77kn\" (UniqueName: \"kubernetes.io/projected/cb09fa9f-7359-46a9-baf8-758474e12535-kube-api-access-s77kn\") pod \"keystone-db-create-bqd4v\" (UID: \"cb09fa9f-7359-46a9-baf8-758474e12535\") " pod="openstack/keystone-db-create-bqd4v" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.302653 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bqd4v" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.326365 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76fc7\" (UniqueName: \"kubernetes.io/projected/e56670c7-7dcd-4294-b852-cf1397b892f7-kube-api-access-76fc7\") pod \"placement-db-create-wc7vd\" (UID: \"e56670c7-7dcd-4294-b852-cf1397b892f7\") " pod="openstack/placement-db-create-wc7vd" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.347687 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76fc7\" (UniqueName: \"kubernetes.io/projected/e56670c7-7dcd-4294-b852-cf1397b892f7-kube-api-access-76fc7\") pod \"placement-db-create-wc7vd\" (UID: \"e56670c7-7dcd-4294-b852-cf1397b892f7\") " pod="openstack/placement-db-create-wc7vd" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.466958 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-d5gmk"] Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.470880 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d5gmk" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.474137 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-d5gmk"] Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.479413 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wc7vd" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.529018 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f45vl\" (UniqueName: \"kubernetes.io/projected/ed3ec4f1-c604-42e1-bffd-255bf7586994-kube-api-access-f45vl\") pod \"glance-db-create-d5gmk\" (UID: \"ed3ec4f1-c604-42e1-bffd-255bf7586994\") " pod="openstack/glance-db-create-d5gmk" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.630315 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f45vl\" (UniqueName: \"kubernetes.io/projected/ed3ec4f1-c604-42e1-bffd-255bf7586994-kube-api-access-f45vl\") pod \"glance-db-create-d5gmk\" (UID: \"ed3ec4f1-c604-42e1-bffd-255bf7586994\") " pod="openstack/glance-db-create-d5gmk" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.649306 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f45vl\" (UniqueName: \"kubernetes.io/projected/ed3ec4f1-c604-42e1-bffd-255bf7586994-kube-api-access-f45vl\") pod \"glance-db-create-d5gmk\" (UID: \"ed3ec4f1-c604-42e1-bffd-255bf7586994\") " pod="openstack/glance-db-create-d5gmk" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.733913 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bqd4v"] Oct 02 09:43:57 crc kubenswrapper[5035]: W1002 09:43:57.745157 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb09fa9f_7359_46a9_baf8_758474e12535.slice/crio-1136e7ac8bfdd5a28a842de0bfd7d1d7bdc822f4c51a33defb400d1b660849df WatchSource:0}: Error finding container 1136e7ac8bfdd5a28a842de0bfd7d1d7bdc822f4c51a33defb400d1b660849df: Status 404 returned error can't find the container with id 1136e7ac8bfdd5a28a842de0bfd7d1d7bdc822f4c51a33defb400d1b660849df Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.775675 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bqd4v" event={"ID":"cb09fa9f-7359-46a9-baf8-758474e12535","Type":"ContainerStarted","Data":"1136e7ac8bfdd5a28a842de0bfd7d1d7bdc822f4c51a33defb400d1b660849df"} Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.797838 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d5gmk" Oct 02 09:43:57 crc kubenswrapper[5035]: I1002 09:43:57.896812 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wc7vd"] Oct 02 09:43:57 crc kubenswrapper[5035]: W1002 09:43:57.912836 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode56670c7_7dcd_4294_b852_cf1397b892f7.slice/crio-bd675da6839443d02272e010070b3c590da759705bd4681a357ee6713355b407 WatchSource:0}: Error finding container bd675da6839443d02272e010070b3c590da759705bd4681a357ee6713355b407: Status 404 returned error can't find the container with id bd675da6839443d02272e010070b3c590da759705bd4681a357ee6713355b407 Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.216267 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-d5gmk"] Oct 02 09:43:58 crc kubenswrapper[5035]: W1002 09:43:58.222478 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded3ec4f1_c604_42e1_bffd_255bf7586994.slice/crio-8cd056e0b24d2a89c88a739c6b81c991b89c243332f28101725763cec8bd63a0 WatchSource:0}: Error finding container 8cd056e0b24d2a89c88a739c6b81c991b89c243332f28101725763cec8bd63a0: Status 404 returned error can't find the container with id 8cd056e0b24d2a89c88a739c6b81c991b89c243332f28101725763cec8bd63a0 Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.545742 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:43:58 crc kubenswrapper[5035]: E1002 09:43:58.545901 5035 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 09:43:58 crc kubenswrapper[5035]: E1002 09:43:58.546092 5035 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 09:43:58 crc kubenswrapper[5035]: E1002 09:43:58.546141 5035 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift podName:2aaa2a11-2896-4690-819b-61c6069ea6db nodeName:}" failed. No retries permitted until 2025-10-02 09:44:06.546126687 +0000 UTC m=+1011.902470712 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift") pod "swift-storage-0" (UID: "2aaa2a11-2896-4690-819b-61c6069ea6db") : configmap "swift-ring-files" not found Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.790608 5035 generic.go:334] "Generic (PLEG): container finished" podID="e56670c7-7dcd-4294-b852-cf1397b892f7" containerID="e547550edd7974f43c6afa0784465d4329e93f5ef3b963dcfd3381d05c8cb767" exitCode=0 Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.790710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wc7vd" event={"ID":"e56670c7-7dcd-4294-b852-cf1397b892f7","Type":"ContainerDied","Data":"e547550edd7974f43c6afa0784465d4329e93f5ef3b963dcfd3381d05c8cb767"} Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.790752 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wc7vd" event={"ID":"e56670c7-7dcd-4294-b852-cf1397b892f7","Type":"ContainerStarted","Data":"bd675da6839443d02272e010070b3c590da759705bd4681a357ee6713355b407"} Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.801519 5035 generic.go:334] "Generic (PLEG): container finished" podID="ed3ec4f1-c604-42e1-bffd-255bf7586994" containerID="0d1f9166aab340a416355f48608eb5a3413a9049b66064a4ba94e116a83aa557" exitCode=0 Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.801628 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d5gmk" event={"ID":"ed3ec4f1-c604-42e1-bffd-255bf7586994","Type":"ContainerDied","Data":"0d1f9166aab340a416355f48608eb5a3413a9049b66064a4ba94e116a83aa557"} Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.801665 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d5gmk" event={"ID":"ed3ec4f1-c604-42e1-bffd-255bf7586994","Type":"ContainerStarted","Data":"8cd056e0b24d2a89c88a739c6b81c991b89c243332f28101725763cec8bd63a0"} Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.807354 5035 generic.go:334] "Generic (PLEG): container finished" podID="cb09fa9f-7359-46a9-baf8-758474e12535" containerID="7ffeb550cf987e795f5d2e0401dbb2dd58531bdaf6b5a851e9b4618f6ee17357" exitCode=0 Oct 02 09:43:58 crc kubenswrapper[5035]: I1002 09:43:58.807435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bqd4v" event={"ID":"cb09fa9f-7359-46a9-baf8-758474e12535","Type":"ContainerDied","Data":"7ffeb550cf987e795f5d2e0401dbb2dd58531bdaf6b5a851e9b4618f6ee17357"} Oct 02 09:43:59 crc kubenswrapper[5035]: I1002 09:43:59.885658 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:43:59 crc kubenswrapper[5035]: I1002 09:43:59.944836 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2ttrv"] Oct 02 09:43:59 crc kubenswrapper[5035]: I1002 09:43:59.945110 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerName="dnsmasq-dns" containerID="cri-o://5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475" gracePeriod=10 Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.261006 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d5gmk" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.265470 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bqd4v" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.286177 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wc7vd" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.379357 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f45vl\" (UniqueName: \"kubernetes.io/projected/ed3ec4f1-c604-42e1-bffd-255bf7586994-kube-api-access-f45vl\") pod \"ed3ec4f1-c604-42e1-bffd-255bf7586994\" (UID: \"ed3ec4f1-c604-42e1-bffd-255bf7586994\") " Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.379591 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s77kn\" (UniqueName: \"kubernetes.io/projected/cb09fa9f-7359-46a9-baf8-758474e12535-kube-api-access-s77kn\") pod \"cb09fa9f-7359-46a9-baf8-758474e12535\" (UID: \"cb09fa9f-7359-46a9-baf8-758474e12535\") " Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.387886 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed3ec4f1-c604-42e1-bffd-255bf7586994-kube-api-access-f45vl" (OuterVolumeSpecName: "kube-api-access-f45vl") pod "ed3ec4f1-c604-42e1-bffd-255bf7586994" (UID: "ed3ec4f1-c604-42e1-bffd-255bf7586994"). InnerVolumeSpecName "kube-api-access-f45vl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.391261 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb09fa9f-7359-46a9-baf8-758474e12535-kube-api-access-s77kn" (OuterVolumeSpecName: "kube-api-access-s77kn") pod "cb09fa9f-7359-46a9-baf8-758474e12535" (UID: "cb09fa9f-7359-46a9-baf8-758474e12535"). InnerVolumeSpecName "kube-api-access-s77kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.433853 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.480961 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76fc7\" (UniqueName: \"kubernetes.io/projected/e56670c7-7dcd-4294-b852-cf1397b892f7-kube-api-access-76fc7\") pod \"e56670c7-7dcd-4294-b852-cf1397b892f7\" (UID: \"e56670c7-7dcd-4294-b852-cf1397b892f7\") " Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.481371 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f45vl\" (UniqueName: \"kubernetes.io/projected/ed3ec4f1-c604-42e1-bffd-255bf7586994-kube-api-access-f45vl\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.481388 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s77kn\" (UniqueName: \"kubernetes.io/projected/cb09fa9f-7359-46a9-baf8-758474e12535-kube-api-access-s77kn\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.485567 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e56670c7-7dcd-4294-b852-cf1397b892f7-kube-api-access-76fc7" (OuterVolumeSpecName: "kube-api-access-76fc7") pod "e56670c7-7dcd-4294-b852-cf1397b892f7" (UID: "e56670c7-7dcd-4294-b852-cf1397b892f7"). InnerVolumeSpecName "kube-api-access-76fc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.582841 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-config\") pod \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.582958 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-dns-svc\") pod \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.583088 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qflg\" (UniqueName: \"kubernetes.io/projected/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-kube-api-access-7qflg\") pod \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\" (UID: \"c4138bce-f108-4a2c-b2fa-8a44f264ccc7\") " Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.583585 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76fc7\" (UniqueName: \"kubernetes.io/projected/e56670c7-7dcd-4294-b852-cf1397b892f7-kube-api-access-76fc7\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.590830 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-kube-api-access-7qflg" (OuterVolumeSpecName: "kube-api-access-7qflg") pod "c4138bce-f108-4a2c-b2fa-8a44f264ccc7" (UID: "c4138bce-f108-4a2c-b2fa-8a44f264ccc7"). InnerVolumeSpecName "kube-api-access-7qflg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.626506 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-config" (OuterVolumeSpecName: "config") pod "c4138bce-f108-4a2c-b2fa-8a44f264ccc7" (UID: "c4138bce-f108-4a2c-b2fa-8a44f264ccc7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.631285 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c4138bce-f108-4a2c-b2fa-8a44f264ccc7" (UID: "c4138bce-f108-4a2c-b2fa-8a44f264ccc7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.685408 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.685453 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qflg\" (UniqueName: \"kubernetes.io/projected/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-kube-api-access-7qflg\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.685471 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4138bce-f108-4a2c-b2fa-8a44f264ccc7-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.824057 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wc7vd" event={"ID":"e56670c7-7dcd-4294-b852-cf1397b892f7","Type":"ContainerDied","Data":"bd675da6839443d02272e010070b3c590da759705bd4681a357ee6713355b407"} Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.824104 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd675da6839443d02272e010070b3c590da759705bd4681a357ee6713355b407" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.824364 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wc7vd" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.825926 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-d5gmk" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.825908 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-d5gmk" event={"ID":"ed3ec4f1-c604-42e1-bffd-255bf7586994","Type":"ContainerDied","Data":"8cd056e0b24d2a89c88a739c6b81c991b89c243332f28101725763cec8bd63a0"} Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.826047 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cd056e0b24d2a89c88a739c6b81c991b89c243332f28101725763cec8bd63a0" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.827692 5035 generic.go:334] "Generic (PLEG): container finished" podID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerID="5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475" exitCode=0 Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.827740 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.827755 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" event={"ID":"c4138bce-f108-4a2c-b2fa-8a44f264ccc7","Type":"ContainerDied","Data":"5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475"} Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.827776 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-2ttrv" event={"ID":"c4138bce-f108-4a2c-b2fa-8a44f264ccc7","Type":"ContainerDied","Data":"624b1a942bfc1508fd5194c9695c9db1e356274572c62dae0757cc07d8877b60"} Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.827793 5035 scope.go:117] "RemoveContainer" containerID="5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.829546 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bqd4v" event={"ID":"cb09fa9f-7359-46a9-baf8-758474e12535","Type":"ContainerDied","Data":"1136e7ac8bfdd5a28a842de0bfd7d1d7bdc822f4c51a33defb400d1b660849df"} Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.829575 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1136e7ac8bfdd5a28a842de0bfd7d1d7bdc822f4c51a33defb400d1b660849df" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.829581 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bqd4v" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.851119 5035 scope.go:117] "RemoveContainer" containerID="6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.879198 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2ttrv"] Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.882335 5035 scope.go:117] "RemoveContainer" containerID="5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475" Oct 02 09:44:00 crc kubenswrapper[5035]: E1002 09:44:00.882850 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475\": container with ID starting with 5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475 not found: ID does not exist" containerID="5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.882887 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475"} err="failed to get container status \"5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475\": rpc error: code = NotFound desc = could not find container \"5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475\": container with ID starting with 5af71b17597a1ff4ab2dfb0cbd587651f8cd85bde596b305ea5a76a0db07e475 not found: ID does not exist" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.882950 5035 scope.go:117] "RemoveContainer" containerID="6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8" Oct 02 09:44:00 crc kubenswrapper[5035]: E1002 09:44:00.883247 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8\": container with ID starting with 6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8 not found: ID does not exist" containerID="6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.883292 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8"} err="failed to get container status \"6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8\": rpc error: code = NotFound desc = could not find container \"6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8\": container with ID starting with 6879825a10bef0b1bb57ddbea11aa8d71bcbdd3c5735fe5bee7bd49cec8f8fe8 not found: ID does not exist" Oct 02 09:44:00 crc kubenswrapper[5035]: I1002 09:44:00.885523 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-2ttrv"] Oct 02 09:44:02 crc kubenswrapper[5035]: I1002 09:44:02.173949 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" path="/var/lib/kubelet/pods/c4138bce-f108-4a2c-b2fa-8a44f264ccc7/volumes" Oct 02 09:44:02 crc kubenswrapper[5035]: I1002 09:44:02.855180 5035 generic.go:334] "Generic (PLEG): container finished" podID="d93ca48d-a2a5-4036-930d-948d935cce98" containerID="e36fa1d3fccf3dedee1297a6e22e93d2b045eb47fc9e1ce5249d3fb5da8dd15e" exitCode=0 Oct 02 09:44:02 crc kubenswrapper[5035]: I1002 09:44:02.855235 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d93ca48d-a2a5-4036-930d-948d935cce98","Type":"ContainerDied","Data":"e36fa1d3fccf3dedee1297a6e22e93d2b045eb47fc9e1ce5249d3fb5da8dd15e"} Oct 02 09:44:03 crc kubenswrapper[5035]: I1002 09:44:03.865614 5035 generic.go:334] "Generic (PLEG): container finished" podID="65e1909b-5dc0-45ad-af61-f3cd8ce48ace" containerID="b9967648c3c14b81f9899758a17798f6e0ac62004727202f37d54af50f16b134" exitCode=0 Oct 02 09:44:03 crc kubenswrapper[5035]: I1002 09:44:03.865733 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wcbql" event={"ID":"65e1909b-5dc0-45ad-af61-f3cd8ce48ace","Type":"ContainerDied","Data":"b9967648c3c14b81f9899758a17798f6e0ac62004727202f37d54af50f16b134"} Oct 02 09:44:03 crc kubenswrapper[5035]: I1002 09:44:03.868486 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d93ca48d-a2a5-4036-930d-948d935cce98","Type":"ContainerStarted","Data":"92fac5eef417f7406dd65bf740166169cb4a298136a1a656499442bebaaf24f3"} Oct 02 09:44:03 crc kubenswrapper[5035]: I1002 09:44:03.868785 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 09:44:03 crc kubenswrapper[5035]: I1002 09:44:03.919431 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.618540573 podStartE2EDuration="51.919414942s" podCreationTimestamp="2025-10-02 09:43:12 +0000 UTC" firstStartedPulling="2025-10-02 09:43:14.745789227 +0000 UTC m=+960.102133252" lastFinishedPulling="2025-10-02 09:43:29.046663596 +0000 UTC m=+974.403007621" observedRunningTime="2025-10-02 09:44:03.915198099 +0000 UTC m=+1009.271542124" watchObservedRunningTime="2025-10-02 09:44:03.919414942 +0000 UTC m=+1009.275758967" Oct 02 09:44:04 crc kubenswrapper[5035]: I1002 09:44:04.355735 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.197669 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.364730 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-dispersionconf\") pod \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.364767 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-combined-ca-bundle\") pod \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.364800 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-etc-swift\") pod \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.364949 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sksfz\" (UniqueName: \"kubernetes.io/projected/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-kube-api-access-sksfz\") pod \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.364993 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-ring-data-devices\") pod \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.365010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-swiftconf\") pod \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.365025 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-scripts\") pod \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\" (UID: \"65e1909b-5dc0-45ad-af61-f3cd8ce48ace\") " Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.366616 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "65e1909b-5dc0-45ad-af61-f3cd8ce48ace" (UID: "65e1909b-5dc0-45ad-af61-f3cd8ce48ace"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.367494 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "65e1909b-5dc0-45ad-af61-f3cd8ce48ace" (UID: "65e1909b-5dc0-45ad-af61-f3cd8ce48ace"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.373408 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "65e1909b-5dc0-45ad-af61-f3cd8ce48ace" (UID: "65e1909b-5dc0-45ad-af61-f3cd8ce48ace"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.388061 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-kube-api-access-sksfz" (OuterVolumeSpecName: "kube-api-access-sksfz") pod "65e1909b-5dc0-45ad-af61-f3cd8ce48ace" (UID: "65e1909b-5dc0-45ad-af61-f3cd8ce48ace"). InnerVolumeSpecName "kube-api-access-sksfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.391546 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-scripts" (OuterVolumeSpecName: "scripts") pod "65e1909b-5dc0-45ad-af61-f3cd8ce48ace" (UID: "65e1909b-5dc0-45ad-af61-f3cd8ce48ace"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.401731 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "65e1909b-5dc0-45ad-af61-f3cd8ce48ace" (UID: "65e1909b-5dc0-45ad-af61-f3cd8ce48ace"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.408106 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65e1909b-5dc0-45ad-af61-f3cd8ce48ace" (UID: "65e1909b-5dc0-45ad-af61-f3cd8ce48ace"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.467093 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sksfz\" (UniqueName: \"kubernetes.io/projected/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-kube-api-access-sksfz\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.467133 5035 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.467149 5035 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.467161 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.467173 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.467184 5035 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.467193 5035 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/65e1909b-5dc0-45ad-af61-f3cd8ce48ace-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.882505 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wcbql" event={"ID":"65e1909b-5dc0-45ad-af61-f3cd8ce48ace","Type":"ContainerDied","Data":"997822288cf8a9f6ec7535d1aa60c200ef29e0d536d366e90bdb392ff907fa37"} Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.882776 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="997822288cf8a9f6ec7535d1aa60c200ef29e0d536d366e90bdb392ff907fa37" Oct 02 09:44:05 crc kubenswrapper[5035]: I1002 09:44:05.882600 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcbql" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.582923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.598471 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2aaa2a11-2896-4690-819b-61c6069ea6db-etc-swift\") pod \"swift-storage-0\" (UID: \"2aaa2a11-2896-4690-819b-61c6069ea6db\") " pod="openstack/swift-storage-0" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.603055 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.997704 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2a86-account-create-rvm4s"] Oct 02 09:44:06 crc kubenswrapper[5035]: E1002 09:44:06.998565 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb09fa9f-7359-46a9-baf8-758474e12535" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998587 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb09fa9f-7359-46a9-baf8-758474e12535" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: E1002 09:44:06.998599 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e56670c7-7dcd-4294-b852-cf1397b892f7" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998607 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e56670c7-7dcd-4294-b852-cf1397b892f7" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: E1002 09:44:06.998620 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed3ec4f1-c604-42e1-bffd-255bf7586994" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998626 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed3ec4f1-c604-42e1-bffd-255bf7586994" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: E1002 09:44:06.998636 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e1909b-5dc0-45ad-af61-f3cd8ce48ace" containerName="swift-ring-rebalance" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998642 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e1909b-5dc0-45ad-af61-f3cd8ce48ace" containerName="swift-ring-rebalance" Oct 02 09:44:06 crc kubenswrapper[5035]: E1002 09:44:06.998652 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerName="init" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998659 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerName="init" Oct 02 09:44:06 crc kubenswrapper[5035]: E1002 09:44:06.998668 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerName="dnsmasq-dns" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998674 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerName="dnsmasq-dns" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998894 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb09fa9f-7359-46a9-baf8-758474e12535" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998907 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4138bce-f108-4a2c-b2fa-8a44f264ccc7" containerName="dnsmasq-dns" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998916 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed3ec4f1-c604-42e1-bffd-255bf7586994" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998928 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e1909b-5dc0-45ad-af61-f3cd8ce48ace" containerName="swift-ring-rebalance" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.998939 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e56670c7-7dcd-4294-b852-cf1397b892f7" containerName="mariadb-database-create" Oct 02 09:44:06 crc kubenswrapper[5035]: I1002 09:44:06.999469 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2a86-account-create-rvm4s" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.002125 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.006504 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2a86-account-create-rvm4s"] Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.091149 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnhgf\" (UniqueName: \"kubernetes.io/projected/533bb1e5-5f6a-4f87-a671-1af95bb013c4-kube-api-access-bnhgf\") pod \"keystone-2a86-account-create-rvm4s\" (UID: \"533bb1e5-5f6a-4f87-a671-1af95bb013c4\") " pod="openstack/keystone-2a86-account-create-rvm4s" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.140818 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 09:44:07 crc kubenswrapper[5035]: W1002 09:44:07.155735 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2aaa2a11_2896_4690_819b_61c6069ea6db.slice/crio-4e13015d228388c90b7e6868fcc4759853cf79e4a4714ca7d4190802c8b2d9c6 WatchSource:0}: Error finding container 4e13015d228388c90b7e6868fcc4759853cf79e4a4714ca7d4190802c8b2d9c6: Status 404 returned error can't find the container with id 4e13015d228388c90b7e6868fcc4759853cf79e4a4714ca7d4190802c8b2d9c6 Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.192763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnhgf\" (UniqueName: \"kubernetes.io/projected/533bb1e5-5f6a-4f87-a671-1af95bb013c4-kube-api-access-bnhgf\") pod \"keystone-2a86-account-create-rvm4s\" (UID: \"533bb1e5-5f6a-4f87-a671-1af95bb013c4\") " pod="openstack/keystone-2a86-account-create-rvm4s" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.215776 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnhgf\" (UniqueName: \"kubernetes.io/projected/533bb1e5-5f6a-4f87-a671-1af95bb013c4-kube-api-access-bnhgf\") pod \"keystone-2a86-account-create-rvm4s\" (UID: \"533bb1e5-5f6a-4f87-a671-1af95bb013c4\") " pod="openstack/keystone-2a86-account-create-rvm4s" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.299241 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1d44-account-create-tgq25"] Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.300446 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1d44-account-create-tgq25" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.302440 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.308805 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1d44-account-create-tgq25"] Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.321918 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2a86-account-create-rvm4s" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.396428 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv6zb\" (UniqueName: \"kubernetes.io/projected/8fbf41c1-35f8-4689-92ce-5338a740e524-kube-api-access-hv6zb\") pod \"placement-1d44-account-create-tgq25\" (UID: \"8fbf41c1-35f8-4689-92ce-5338a740e524\") " pod="openstack/placement-1d44-account-create-tgq25" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.498361 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv6zb\" (UniqueName: \"kubernetes.io/projected/8fbf41c1-35f8-4689-92ce-5338a740e524-kube-api-access-hv6zb\") pod \"placement-1d44-account-create-tgq25\" (UID: \"8fbf41c1-35f8-4689-92ce-5338a740e524\") " pod="openstack/placement-1d44-account-create-tgq25" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.516115 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv6zb\" (UniqueName: \"kubernetes.io/projected/8fbf41c1-35f8-4689-92ce-5338a740e524-kube-api-access-hv6zb\") pod \"placement-1d44-account-create-tgq25\" (UID: \"8fbf41c1-35f8-4689-92ce-5338a740e524\") " pod="openstack/placement-1d44-account-create-tgq25" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.623957 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1d44-account-create-tgq25" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.644094 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-11a0-account-create-xf5dg"] Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.645029 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11a0-account-create-xf5dg" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.646831 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.655498 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-11a0-account-create-xf5dg"] Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.734213 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2a86-account-create-rvm4s"] Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.808929 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9knq7\" (UniqueName: \"kubernetes.io/projected/772fa808-a60d-4cbf-a6a8-b7e7e2345ad1-kube-api-access-9knq7\") pod \"glance-11a0-account-create-xf5dg\" (UID: \"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1\") " pod="openstack/glance-11a0-account-create-xf5dg" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.868064 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1d44-account-create-tgq25"] Oct 02 09:44:07 crc kubenswrapper[5035]: W1002 09:44:07.873146 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fbf41c1_35f8_4689_92ce_5338a740e524.slice/crio-e538100257efb3376644151414be50730204366fce999cbc223384fbb2e94259 WatchSource:0}: Error finding container e538100257efb3376644151414be50730204366fce999cbc223384fbb2e94259: Status 404 returned error can't find the container with id e538100257efb3376644151414be50730204366fce999cbc223384fbb2e94259 Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.896311 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1d44-account-create-tgq25" event={"ID":"8fbf41c1-35f8-4689-92ce-5338a740e524","Type":"ContainerStarted","Data":"e538100257efb3376644151414be50730204366fce999cbc223384fbb2e94259"} Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.897329 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2a86-account-create-rvm4s" event={"ID":"533bb1e5-5f6a-4f87-a671-1af95bb013c4","Type":"ContainerStarted","Data":"703e9f2c5b6ebf226ebddcc36859a54f182c0a633bbff31e2df0a983a5d74897"} Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.898386 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"4e13015d228388c90b7e6868fcc4759853cf79e4a4714ca7d4190802c8b2d9c6"} Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.910815 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9knq7\" (UniqueName: \"kubernetes.io/projected/772fa808-a60d-4cbf-a6a8-b7e7e2345ad1-kube-api-access-9knq7\") pod \"glance-11a0-account-create-xf5dg\" (UID: \"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1\") " pod="openstack/glance-11a0-account-create-xf5dg" Oct 02 09:44:07 crc kubenswrapper[5035]: I1002 09:44:07.935284 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9knq7\" (UniqueName: \"kubernetes.io/projected/772fa808-a60d-4cbf-a6a8-b7e7e2345ad1-kube-api-access-9knq7\") pod \"glance-11a0-account-create-xf5dg\" (UID: \"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1\") " pod="openstack/glance-11a0-account-create-xf5dg" Oct 02 09:44:08 crc kubenswrapper[5035]: I1002 09:44:08.080456 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11a0-account-create-xf5dg" Oct 02 09:44:08 crc kubenswrapper[5035]: I1002 09:44:08.516813 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-11a0-account-create-xf5dg"] Oct 02 09:44:08 crc kubenswrapper[5035]: I1002 09:44:08.909775 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11a0-account-create-xf5dg" event={"ID":"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1","Type":"ContainerStarted","Data":"f0d1236fe1f8d7b2cd8f5d16b0ac1919f0e722668bfab3270f8350c8c4905ed7"} Oct 02 09:44:10 crc kubenswrapper[5035]: I1002 09:44:10.927756 5035 generic.go:334] "Generic (PLEG): container finished" podID="533bb1e5-5f6a-4f87-a671-1af95bb013c4" containerID="27a080b5795472921ae4e21f43bb5869372bb174c20776b899ade1b0191e9cc2" exitCode=0 Oct 02 09:44:10 crc kubenswrapper[5035]: I1002 09:44:10.927843 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2a86-account-create-rvm4s" event={"ID":"533bb1e5-5f6a-4f87-a671-1af95bb013c4","Type":"ContainerDied","Data":"27a080b5795472921ae4e21f43bb5869372bb174c20776b899ade1b0191e9cc2"} Oct 02 09:44:10 crc kubenswrapper[5035]: I1002 09:44:10.930293 5035 generic.go:334] "Generic (PLEG): container finished" podID="772fa808-a60d-4cbf-a6a8-b7e7e2345ad1" containerID="d26dbb8808e679c5fa7c9faac4b005d667e058c6c1f127995e662021738fa8bf" exitCode=0 Oct 02 09:44:10 crc kubenswrapper[5035]: I1002 09:44:10.930389 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11a0-account-create-xf5dg" event={"ID":"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1","Type":"ContainerDied","Data":"d26dbb8808e679c5fa7c9faac4b005d667e058c6c1f127995e662021738fa8bf"} Oct 02 09:44:10 crc kubenswrapper[5035]: I1002 09:44:10.931916 5035 generic.go:334] "Generic (PLEG): container finished" podID="8fbf41c1-35f8-4689-92ce-5338a740e524" containerID="c971a4caed2be13cdbd97fc3bcd7270df0f68140db6230049b17ea142569a286" exitCode=0 Oct 02 09:44:10 crc kubenswrapper[5035]: I1002 09:44:10.931953 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1d44-account-create-tgq25" event={"ID":"8fbf41c1-35f8-4689-92ce-5338a740e524","Type":"ContainerDied","Data":"c971a4caed2be13cdbd97fc3bcd7270df0f68140db6230049b17ea142569a286"} Oct 02 09:44:11 crc kubenswrapper[5035]: I1002 09:44:11.944331 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"85d7a8e6d90945787625181cef05b59c041bea8b83460583084066376d17c251"} Oct 02 09:44:11 crc kubenswrapper[5035]: I1002 09:44:11.944925 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"620b877bf01a7122e22466e7d6046526640d9c82b4bd8ad5458c25fd74a1cb93"} Oct 02 09:44:11 crc kubenswrapper[5035]: I1002 09:44:11.944941 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"476629a30e5e1e2c8b3e754b0d2ebf949db0f3fc97599493f4becaba221457c1"} Oct 02 09:44:11 crc kubenswrapper[5035]: I1002 09:44:11.944949 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"72bcbb2d4d2ef4e3336883d225d24a3b24cc6811046c34f5fd8b5a388e58d5e4"} Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.401972 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jvdts" podUID="57ea64fd-e931-45ab-9f0e-678c3c9ba7a2" containerName="ovn-controller" probeResult="failure" output=< Oct 02 09:44:12 crc kubenswrapper[5035]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 09:44:12 crc kubenswrapper[5035]: > Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.402340 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2a86-account-create-rvm4s" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.410393 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.413042 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1d44-account-create-tgq25" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.415794 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-r5m97" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.433091 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11a0-account-create-xf5dg" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.500252 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9knq7\" (UniqueName: \"kubernetes.io/projected/772fa808-a60d-4cbf-a6a8-b7e7e2345ad1-kube-api-access-9knq7\") pod \"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1\" (UID: \"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1\") " Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.500509 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnhgf\" (UniqueName: \"kubernetes.io/projected/533bb1e5-5f6a-4f87-a671-1af95bb013c4-kube-api-access-bnhgf\") pod \"533bb1e5-5f6a-4f87-a671-1af95bb013c4\" (UID: \"533bb1e5-5f6a-4f87-a671-1af95bb013c4\") " Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.500569 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv6zb\" (UniqueName: \"kubernetes.io/projected/8fbf41c1-35f8-4689-92ce-5338a740e524-kube-api-access-hv6zb\") pod \"8fbf41c1-35f8-4689-92ce-5338a740e524\" (UID: \"8fbf41c1-35f8-4689-92ce-5338a740e524\") " Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.507167 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fbf41c1-35f8-4689-92ce-5338a740e524-kube-api-access-hv6zb" (OuterVolumeSpecName: "kube-api-access-hv6zb") pod "8fbf41c1-35f8-4689-92ce-5338a740e524" (UID: "8fbf41c1-35f8-4689-92ce-5338a740e524"). InnerVolumeSpecName "kube-api-access-hv6zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.508677 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/533bb1e5-5f6a-4f87-a671-1af95bb013c4-kube-api-access-bnhgf" (OuterVolumeSpecName: "kube-api-access-bnhgf") pod "533bb1e5-5f6a-4f87-a671-1af95bb013c4" (UID: "533bb1e5-5f6a-4f87-a671-1af95bb013c4"). InnerVolumeSpecName "kube-api-access-bnhgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.509817 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/772fa808-a60d-4cbf-a6a8-b7e7e2345ad1-kube-api-access-9knq7" (OuterVolumeSpecName: "kube-api-access-9knq7") pod "772fa808-a60d-4cbf-a6a8-b7e7e2345ad1" (UID: "772fa808-a60d-4cbf-a6a8-b7e7e2345ad1"). InnerVolumeSpecName "kube-api-access-9knq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.602456 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnhgf\" (UniqueName: \"kubernetes.io/projected/533bb1e5-5f6a-4f87-a671-1af95bb013c4-kube-api-access-bnhgf\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.602557 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv6zb\" (UniqueName: \"kubernetes.io/projected/8fbf41c1-35f8-4689-92ce-5338a740e524-kube-api-access-hv6zb\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.602573 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9knq7\" (UniqueName: \"kubernetes.io/projected/772fa808-a60d-4cbf-a6a8-b7e7e2345ad1-kube-api-access-9knq7\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.665962 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jvdts-config-4h9wh"] Oct 02 09:44:12 crc kubenswrapper[5035]: E1002 09:44:12.666274 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="533bb1e5-5f6a-4f87-a671-1af95bb013c4" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.666290 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="533bb1e5-5f6a-4f87-a671-1af95bb013c4" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: E1002 09:44:12.666318 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772fa808-a60d-4cbf-a6a8-b7e7e2345ad1" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.666324 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="772fa808-a60d-4cbf-a6a8-b7e7e2345ad1" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: E1002 09:44:12.666333 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbf41c1-35f8-4689-92ce-5338a740e524" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.666340 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbf41c1-35f8-4689-92ce-5338a740e524" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.666847 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="533bb1e5-5f6a-4f87-a671-1af95bb013c4" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.666869 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fbf41c1-35f8-4689-92ce-5338a740e524" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.666879 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="772fa808-a60d-4cbf-a6a8-b7e7e2345ad1" containerName="mariadb-account-create" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.667591 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.674467 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.678808 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jvdts-config-4h9wh"] Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.806047 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run-ovn\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.806101 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-scripts\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.806155 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-additional-scripts\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.806228 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.806272 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xkv5\" (UniqueName: \"kubernetes.io/projected/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-kube-api-access-5xkv5\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.806319 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-log-ovn\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.908633 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run-ovn\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.908699 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-scripts\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.908737 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-additional-scripts\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.908803 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.909019 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run-ovn\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.910065 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xkv5\" (UniqueName: \"kubernetes.io/projected/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-kube-api-access-5xkv5\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.910092 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.910267 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-log-ovn\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.910647 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-log-ovn\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.911183 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-additional-scripts\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.912171 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-scripts\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.927361 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xkv5\" (UniqueName: \"kubernetes.io/projected/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-kube-api-access-5xkv5\") pod \"ovn-controller-jvdts-config-4h9wh\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.955329 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1d44-account-create-tgq25" event={"ID":"8fbf41c1-35f8-4689-92ce-5338a740e524","Type":"ContainerDied","Data":"e538100257efb3376644151414be50730204366fce999cbc223384fbb2e94259"} Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.955389 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e538100257efb3376644151414be50730204366fce999cbc223384fbb2e94259" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.955349 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1d44-account-create-tgq25" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.959218 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2a86-account-create-rvm4s" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.959292 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2a86-account-create-rvm4s" event={"ID":"533bb1e5-5f6a-4f87-a671-1af95bb013c4","Type":"ContainerDied","Data":"703e9f2c5b6ebf226ebddcc36859a54f182c0a633bbff31e2df0a983a5d74897"} Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.959339 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="703e9f2c5b6ebf226ebddcc36859a54f182c0a633bbff31e2df0a983a5d74897" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.961292 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11a0-account-create-xf5dg" event={"ID":"772fa808-a60d-4cbf-a6a8-b7e7e2345ad1","Type":"ContainerDied","Data":"f0d1236fe1f8d7b2cd8f5d16b0ac1919f0e722668bfab3270f8350c8c4905ed7"} Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.961340 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0d1236fe1f8d7b2cd8f5d16b0ac1919f0e722668bfab3270f8350c8c4905ed7" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.961360 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11a0-account-create-xf5dg" Oct 02 09:44:12 crc kubenswrapper[5035]: I1002 09:44:12.989719 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:13 crc kubenswrapper[5035]: I1002 09:44:13.535974 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jvdts-config-4h9wh"] Oct 02 09:44:13 crc kubenswrapper[5035]: W1002 09:44:13.543048 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84b6ce68_ccb0_40ab_a4b6_7cf7ee7423ff.slice/crio-54839fc481cd35a1767613aeb5d2ba27298475e747706414a5577342969ae8ac WatchSource:0}: Error finding container 54839fc481cd35a1767613aeb5d2ba27298475e747706414a5577342969ae8ac: Status 404 returned error can't find the container with id 54839fc481cd35a1767613aeb5d2ba27298475e747706414a5577342969ae8ac Oct 02 09:44:13 crc kubenswrapper[5035]: I1002 09:44:13.976702 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"f66b3c556bfabd79951790cd720564aa0ded49a111dc3c6d184270947292c8c8"} Oct 02 09:44:13 crc kubenswrapper[5035]: I1002 09:44:13.979605 5035 generic.go:334] "Generic (PLEG): container finished" podID="3d7eafce-151f-43f3-8f74-cbfc9f368643" containerID="f1a518f45103e2a6dd09b8cc44fd94b3fc5bcbbd39842e6c0d7bc2c73063a1f3" exitCode=0 Oct 02 09:44:13 crc kubenswrapper[5035]: I1002 09:44:13.979724 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3d7eafce-151f-43f3-8f74-cbfc9f368643","Type":"ContainerDied","Data":"f1a518f45103e2a6dd09b8cc44fd94b3fc5bcbbd39842e6c0d7bc2c73063a1f3"} Oct 02 09:44:13 crc kubenswrapper[5035]: I1002 09:44:13.983468 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts-config-4h9wh" event={"ID":"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff","Type":"ContainerStarted","Data":"03f45f06468cc4881431ccb308d8a34e6a4960701f3592cf3dcf729cf1921bdf"} Oct 02 09:44:13 crc kubenswrapper[5035]: I1002 09:44:13.983573 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts-config-4h9wh" event={"ID":"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff","Type":"ContainerStarted","Data":"54839fc481cd35a1767613aeb5d2ba27298475e747706414a5577342969ae8ac"} Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.032071 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jvdts-config-4h9wh" podStartSLOduration=2.03205112 podStartE2EDuration="2.03205112s" podCreationTimestamp="2025-10-02 09:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:14.022122501 +0000 UTC m=+1019.378466526" watchObservedRunningTime="2025-10-02 09:44:14.03205112 +0000 UTC m=+1019.388395145" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.100761 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.441218 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-jmk82"] Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.442587 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jmk82" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.465264 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jmk82"] Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.545754 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7cdk\" (UniqueName: \"kubernetes.io/projected/76e148c5-87f2-440a-adae-0d3cc1f67f2a-kube-api-access-h7cdk\") pod \"cinder-db-create-jmk82\" (UID: \"76e148c5-87f2-440a-adae-0d3cc1f67f2a\") " pod="openstack/cinder-db-create-jmk82" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.560246 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-tkjv5"] Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.561614 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tkjv5" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.581257 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tkjv5"] Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.647184 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7cdk\" (UniqueName: \"kubernetes.io/projected/76e148c5-87f2-440a-adae-0d3cc1f67f2a-kube-api-access-h7cdk\") pod \"cinder-db-create-jmk82\" (UID: \"76e148c5-87f2-440a-adae-0d3cc1f67f2a\") " pod="openstack/cinder-db-create-jmk82" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.647258 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6np5\" (UniqueName: \"kubernetes.io/projected/68687a35-da55-497b-a7fa-0fe15b039d15-kube-api-access-k6np5\") pod \"barbican-db-create-tkjv5\" (UID: \"68687a35-da55-497b-a7fa-0fe15b039d15\") " pod="openstack/barbican-db-create-tkjv5" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.684647 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7cdk\" (UniqueName: \"kubernetes.io/projected/76e148c5-87f2-440a-adae-0d3cc1f67f2a-kube-api-access-h7cdk\") pod \"cinder-db-create-jmk82\" (UID: \"76e148c5-87f2-440a-adae-0d3cc1f67f2a\") " pod="openstack/cinder-db-create-jmk82" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.748548 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6np5\" (UniqueName: \"kubernetes.io/projected/68687a35-da55-497b-a7fa-0fe15b039d15-kube-api-access-k6np5\") pod \"barbican-db-create-tkjv5\" (UID: \"68687a35-da55-497b-a7fa-0fe15b039d15\") " pod="openstack/barbican-db-create-tkjv5" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.748787 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-gjbv9"] Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.749946 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjbv9" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.759677 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jmk82" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.776065 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6np5\" (UniqueName: \"kubernetes.io/projected/68687a35-da55-497b-a7fa-0fe15b039d15-kube-api-access-k6np5\") pod \"barbican-db-create-tkjv5\" (UID: \"68687a35-da55-497b-a7fa-0fe15b039d15\") " pod="openstack/barbican-db-create-tkjv5" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.786810 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-gjbv9"] Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.850212 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d62cj\" (UniqueName: \"kubernetes.io/projected/e29f134c-e329-4296-b491-101f71a255a0-kube-api-access-d62cj\") pod \"neutron-db-create-gjbv9\" (UID: \"e29f134c-e329-4296-b491-101f71a255a0\") " pod="openstack/neutron-db-create-gjbv9" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.952283 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d62cj\" (UniqueName: \"kubernetes.io/projected/e29f134c-e329-4296-b491-101f71a255a0-kube-api-access-d62cj\") pod \"neutron-db-create-gjbv9\" (UID: \"e29f134c-e329-4296-b491-101f71a255a0\") " pod="openstack/neutron-db-create-gjbv9" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.974008 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d62cj\" (UniqueName: \"kubernetes.io/projected/e29f134c-e329-4296-b491-101f71a255a0-kube-api-access-d62cj\") pod \"neutron-db-create-gjbv9\" (UID: \"e29f134c-e329-4296-b491-101f71a255a0\") " pod="openstack/neutron-db-create-gjbv9" Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.995772 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3d7eafce-151f-43f3-8f74-cbfc9f368643","Type":"ContainerStarted","Data":"af9807799c6e9ecfd20e9aac19e4d6a08eeab45a91301a6c50929f23396438ec"} Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.998274 5035 generic.go:334] "Generic (PLEG): container finished" podID="84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" containerID="03f45f06468cc4881431ccb308d8a34e6a4960701f3592cf3dcf729cf1921bdf" exitCode=0 Oct 02 09:44:14 crc kubenswrapper[5035]: I1002 09:44:14.998403 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts-config-4h9wh" event={"ID":"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff","Type":"ContainerDied","Data":"03f45f06468cc4881431ccb308d8a34e6a4960701f3592cf3dcf729cf1921bdf"} Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.005820 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"b3f4b287ae7a46c36c66bdf38e56eeaf27a3e50e85238664e4a464b219e17dc3"} Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.005874 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"4d6ab8b7064f87de4e04f4f66abb7a18bca5dc7321ddaa85071574e465a84fa5"} Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.005890 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"c915c8ad9cc2a9a26aed1af13a470610f34a393f0110a5f3b12464a42c37a94d"} Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.035962 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=-9223371973.818836 podStartE2EDuration="1m3.035939298s" podCreationTimestamp="2025-10-02 09:43:12 +0000 UTC" firstStartedPulling="2025-10-02 09:43:14.249460648 +0000 UTC m=+959.605804683" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:15.020790587 +0000 UTC m=+1020.377134612" watchObservedRunningTime="2025-10-02 09:44:15.035939298 +0000 UTC m=+1020.392283333" Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.037092 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tkjv5" Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.119306 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjbv9" Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.214323 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jmk82"] Oct 02 09:44:15 crc kubenswrapper[5035]: W1002 09:44:15.223307 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76e148c5_87f2_440a_adae_0d3cc1f67f2a.slice/crio-563b24c18b6bf5f15441c11fe4c90a3cbcda159efd532b65e1de5f8e9b716316 WatchSource:0}: Error finding container 563b24c18b6bf5f15441c11fe4c90a3cbcda159efd532b65e1de5f8e9b716316: Status 404 returned error can't find the container with id 563b24c18b6bf5f15441c11fe4c90a3cbcda159efd532b65e1de5f8e9b716316 Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.572020 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-gjbv9"] Oct 02 09:44:15 crc kubenswrapper[5035]: I1002 09:44:15.596956 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tkjv5"] Oct 02 09:44:15 crc kubenswrapper[5035]: W1002 09:44:15.635488 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68687a35_da55_497b_a7fa_0fe15b039d15.slice/crio-b897c5f7a243e9735b79d27b9603072ef564e1639c66126a433f9b5e3038cbe8 WatchSource:0}: Error finding container b897c5f7a243e9735b79d27b9603072ef564e1639c66126a433f9b5e3038cbe8: Status 404 returned error can't find the container with id b897c5f7a243e9735b79d27b9603072ef564e1639c66126a433f9b5e3038cbe8 Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.015290 5035 generic.go:334] "Generic (PLEG): container finished" podID="76e148c5-87f2-440a-adae-0d3cc1f67f2a" containerID="20dc0a2ea25fa97bb7ae731fe42920cb0582ac2484f3e902bee72485975fcc90" exitCode=0 Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.015391 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jmk82" event={"ID":"76e148c5-87f2-440a-adae-0d3cc1f67f2a","Type":"ContainerDied","Data":"20dc0a2ea25fa97bb7ae731fe42920cb0582ac2484f3e902bee72485975fcc90"} Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.015651 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jmk82" event={"ID":"76e148c5-87f2-440a-adae-0d3cc1f67f2a","Type":"ContainerStarted","Data":"563b24c18b6bf5f15441c11fe4c90a3cbcda159efd532b65e1de5f8e9b716316"} Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.016950 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tkjv5" event={"ID":"68687a35-da55-497b-a7fa-0fe15b039d15","Type":"ContainerStarted","Data":"4653b9a0a9c73a2a75d3f3ade0365b3d1a7f78880a01063517758e0f30840ebc"} Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.017013 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tkjv5" event={"ID":"68687a35-da55-497b-a7fa-0fe15b039d15","Type":"ContainerStarted","Data":"b897c5f7a243e9735b79d27b9603072ef564e1639c66126a433f9b5e3038cbe8"} Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.019082 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gjbv9" event={"ID":"e29f134c-e329-4296-b491-101f71a255a0","Type":"ContainerStarted","Data":"4ccde4beae1e00ad4b185cfebf60366ec2f45f85fa1fb6d555f0521ea26b7c18"} Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.019123 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gjbv9" event={"ID":"e29f134c-e329-4296-b491-101f71a255a0","Type":"ContainerStarted","Data":"a8b644360a5cb31cde46e3f0c7798e85831e1f0496183f4235392cf9ab970f38"} Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.605504 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.691946 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xkv5\" (UniqueName: \"kubernetes.io/projected/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-kube-api-access-5xkv5\") pod \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692118 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-scripts\") pod \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692163 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run-ovn\") pod \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692202 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-log-ovn\") pod \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692263 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-additional-scripts\") pod \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692622 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run\") pod \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\" (UID: \"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff\") " Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692782 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" (UID: "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692847 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" (UID: "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692870 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" (UID: "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.692889 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run" (OuterVolumeSpecName: "var-run") pod "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" (UID: "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.693317 5035 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.693337 5035 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.693346 5035 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.693355 5035 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.693402 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-scripts" (OuterVolumeSpecName: "scripts") pod "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" (UID: "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.703137 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-kube-api-access-5xkv5" (OuterVolumeSpecName: "kube-api-access-5xkv5") pod "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" (UID: "84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff"). InnerVolumeSpecName "kube-api-access-5xkv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.795023 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:16 crc kubenswrapper[5035]: I1002 09:44:16.795344 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xkv5\" (UniqueName: \"kubernetes.io/projected/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff-kube-api-access-5xkv5\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.032818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts-config-4h9wh" event={"ID":"84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff","Type":"ContainerDied","Data":"54839fc481cd35a1767613aeb5d2ba27298475e747706414a5577342969ae8ac"} Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.032865 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54839fc481cd35a1767613aeb5d2ba27298475e747706414a5577342969ae8ac" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.034072 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-4h9wh" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.034932 5035 generic.go:334] "Generic (PLEG): container finished" podID="68687a35-da55-497b-a7fa-0fe15b039d15" containerID="4653b9a0a9c73a2a75d3f3ade0365b3d1a7f78880a01063517758e0f30840ebc" exitCode=0 Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.035352 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tkjv5" event={"ID":"68687a35-da55-497b-a7fa-0fe15b039d15","Type":"ContainerDied","Data":"4653b9a0a9c73a2a75d3f3ade0365b3d1a7f78880a01063517758e0f30840ebc"} Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.037040 5035 generic.go:334] "Generic (PLEG): container finished" podID="e29f134c-e329-4296-b491-101f71a255a0" containerID="4ccde4beae1e00ad4b185cfebf60366ec2f45f85fa1fb6d555f0521ea26b7c18" exitCode=0 Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.037093 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gjbv9" event={"ID":"e29f134c-e329-4296-b491-101f71a255a0","Type":"ContainerDied","Data":"4ccde4beae1e00ad4b185cfebf60366ec2f45f85fa1fb6d555f0521ea26b7c18"} Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.041597 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"e8683316ac7114aa1215423c55252d57f3c2a155696d8d3daacc5ef8b1422f9b"} Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.041631 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"a1de4e5fa017d6f6c68d12c67b100b5d2cc6044e2062cee43820fdfef6283f9f"} Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.168042 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jvdts-config-4h9wh"] Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.176963 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jvdts-config-4h9wh"] Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.297805 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jvdts-config-h96d9"] Oct 02 09:44:17 crc kubenswrapper[5035]: E1002 09:44:17.298186 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" containerName="ovn-config" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.298198 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" containerName="ovn-config" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.298370 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" containerName="ovn-config" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.298925 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.307431 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.314751 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jvdts-config-h96d9"] Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.409716 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run-ovn\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.410173 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-additional-scripts\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.410248 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-log-ovn\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.410318 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.410353 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-scripts\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.410380 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhfcc\" (UniqueName: \"kubernetes.io/projected/97170441-9666-4bb1-856a-a229bca4810d-kube-api-access-dhfcc\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.418071 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jvdts" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.512237 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhfcc\" (UniqueName: \"kubernetes.io/projected/97170441-9666-4bb1-856a-a229bca4810d-kube-api-access-dhfcc\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.513235 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run-ovn\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.513648 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run-ovn\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.514604 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-additional-scripts\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.514830 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-log-ovn\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.514927 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.514952 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-log-ovn\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.515043 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-scripts\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.515076 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.515493 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-additional-scripts\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.520493 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-scripts\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.527647 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jmk82" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.532598 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhfcc\" (UniqueName: \"kubernetes.io/projected/97170441-9666-4bb1-856a-a229bca4810d-kube-api-access-dhfcc\") pod \"ovn-controller-jvdts-config-h96d9\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.616285 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7cdk\" (UniqueName: \"kubernetes.io/projected/76e148c5-87f2-440a-adae-0d3cc1f67f2a-kube-api-access-h7cdk\") pod \"76e148c5-87f2-440a-adae-0d3cc1f67f2a\" (UID: \"76e148c5-87f2-440a-adae-0d3cc1f67f2a\") " Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.619474 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76e148c5-87f2-440a-adae-0d3cc1f67f2a-kube-api-access-h7cdk" (OuterVolumeSpecName: "kube-api-access-h7cdk") pod "76e148c5-87f2-440a-adae-0d3cc1f67f2a" (UID: "76e148c5-87f2-440a-adae-0d3cc1f67f2a"). InnerVolumeSpecName "kube-api-access-h7cdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.686229 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.692754 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-zhwgl"] Oct 02 09:44:17 crc kubenswrapper[5035]: E1002 09:44:17.693370 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76e148c5-87f2-440a-adae-0d3cc1f67f2a" containerName="mariadb-database-create" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.693395 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="76e148c5-87f2-440a-adae-0d3cc1f67f2a" containerName="mariadb-database-create" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.693679 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="76e148c5-87f2-440a-adae-0d3cc1f67f2a" containerName="mariadb-database-create" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.695353 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.702020 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.702330 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fj9kx" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.703835 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.704900 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.712168 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zhwgl"] Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.717968 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7cdk\" (UniqueName: \"kubernetes.io/projected/76e148c5-87f2-440a-adae-0d3cc1f67f2a-kube-api-access-h7cdk\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.819127 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgxqt\" (UniqueName: \"kubernetes.io/projected/5bf48931-dee2-4f51-a3dc-d69a4604064b-kube-api-access-cgxqt\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.819443 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-config-data\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.819517 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-combined-ca-bundle\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.853372 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-hbb7f"] Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.854760 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.857299 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.857727 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2l9tg" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.869255 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-hbb7f"] Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.921608 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-config-data\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.921676 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-config-data\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.921779 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-combined-ca-bundle\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.921811 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-db-sync-config-data\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.921862 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-combined-ca-bundle\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.921904 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfnds\" (UniqueName: \"kubernetes.io/projected/f6f3e489-f6ce-4126-adcb-764afd7e7159-kube-api-access-kfnds\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.921934 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgxqt\" (UniqueName: \"kubernetes.io/projected/5bf48931-dee2-4f51-a3dc-d69a4604064b-kube-api-access-cgxqt\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.929472 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-config-data\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.935220 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-combined-ca-bundle\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:17 crc kubenswrapper[5035]: I1002 09:44:17.940939 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgxqt\" (UniqueName: \"kubernetes.io/projected/5bf48931-dee2-4f51-a3dc-d69a4604064b-kube-api-access-cgxqt\") pod \"keystone-db-sync-zhwgl\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.017509 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.023453 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-db-sync-config-data\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.023543 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-combined-ca-bundle\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.023577 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfnds\" (UniqueName: \"kubernetes.io/projected/f6f3e489-f6ce-4126-adcb-764afd7e7159-kube-api-access-kfnds\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.023631 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-config-data\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.027934 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-combined-ca-bundle\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.028320 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-db-sync-config-data\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.029117 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-config-data\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.045139 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfnds\" (UniqueName: \"kubernetes.io/projected/f6f3e489-f6ce-4126-adcb-764afd7e7159-kube-api-access-kfnds\") pod \"glance-db-sync-hbb7f\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.055039 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jmk82" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.055029 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jmk82" event={"ID":"76e148c5-87f2-440a-adae-0d3cc1f67f2a","Type":"ContainerDied","Data":"563b24c18b6bf5f15441c11fe4c90a3cbcda159efd532b65e1de5f8e9b716316"} Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.055182 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="563b24c18b6bf5f15441c11fe4c90a3cbcda159efd532b65e1de5f8e9b716316" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.083112 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"2016a249813ee9f2d0cc68e497a39a394de880543e198d7b99979a48d2be3f14"} Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.083402 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"9d7d8a5ce8b96d667223fe5fbbc214e735314d6adea1b861d0b97101277c73e9"} Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.083605 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"90de908be3dc0e3fbf9664a5355cb33b9cc407f1878ee5fa43a13cb51ef33bd3"} Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.083687 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"f3de5debd758f51337079b03cdf0188459a78b5bd652528e8c439cd95d884435"} Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.197825 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.214753 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff" path="/var/lib/kubelet/pods/84b6ce68-ccb0-40ab-a4b6-7cf7ee7423ff/volumes" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.216385 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jvdts-config-h96d9"] Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.600233 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjbv9" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.631980 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tkjv5" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.635441 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d62cj\" (UniqueName: \"kubernetes.io/projected/e29f134c-e329-4296-b491-101f71a255a0-kube-api-access-d62cj\") pod \"e29f134c-e329-4296-b491-101f71a255a0\" (UID: \"e29f134c-e329-4296-b491-101f71a255a0\") " Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.642899 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e29f134c-e329-4296-b491-101f71a255a0-kube-api-access-d62cj" (OuterVolumeSpecName: "kube-api-access-d62cj") pod "e29f134c-e329-4296-b491-101f71a255a0" (UID: "e29f134c-e329-4296-b491-101f71a255a0"). InnerVolumeSpecName "kube-api-access-d62cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.722271 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-zhwgl"] Oct 02 09:44:18 crc kubenswrapper[5035]: W1002 09:44:18.731209 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bf48931_dee2_4f51_a3dc_d69a4604064b.slice/crio-a0bb2b212c2c19aac312c2a2cc3b38b89b5d3ff4dc4f49d29b4c24e605b8d5e1 WatchSource:0}: Error finding container a0bb2b212c2c19aac312c2a2cc3b38b89b5d3ff4dc4f49d29b4c24e605b8d5e1: Status 404 returned error can't find the container with id a0bb2b212c2c19aac312c2a2cc3b38b89b5d3ff4dc4f49d29b4c24e605b8d5e1 Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.737268 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6np5\" (UniqueName: \"kubernetes.io/projected/68687a35-da55-497b-a7fa-0fe15b039d15-kube-api-access-k6np5\") pod \"68687a35-da55-497b-a7fa-0fe15b039d15\" (UID: \"68687a35-da55-497b-a7fa-0fe15b039d15\") " Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.737762 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d62cj\" (UniqueName: \"kubernetes.io/projected/e29f134c-e329-4296-b491-101f71a255a0-kube-api-access-d62cj\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.741846 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68687a35-da55-497b-a7fa-0fe15b039d15-kube-api-access-k6np5" (OuterVolumeSpecName: "kube-api-access-k6np5") pod "68687a35-da55-497b-a7fa-0fe15b039d15" (UID: "68687a35-da55-497b-a7fa-0fe15b039d15"). InnerVolumeSpecName "kube-api-access-k6np5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.840405 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6np5\" (UniqueName: \"kubernetes.io/projected/68687a35-da55-497b-a7fa-0fe15b039d15-kube-api-access-k6np5\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:18 crc kubenswrapper[5035]: I1002 09:44:18.900843 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-hbb7f"] Oct 02 09:44:18 crc kubenswrapper[5035]: W1002 09:44:18.903488 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6f3e489_f6ce_4126_adcb_764afd7e7159.slice/crio-3fbf9054afc050643754272d95d1919a8e11528c3ffdc041078fe5861e271a3d WatchSource:0}: Error finding container 3fbf9054afc050643754272d95d1919a8e11528c3ffdc041078fe5861e271a3d: Status 404 returned error can't find the container with id 3fbf9054afc050643754272d95d1919a8e11528c3ffdc041078fe5861e271a3d Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.091664 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhwgl" event={"ID":"5bf48931-dee2-4f51-a3dc-d69a4604064b","Type":"ContainerStarted","Data":"a0bb2b212c2c19aac312c2a2cc3b38b89b5d3ff4dc4f49d29b4c24e605b8d5e1"} Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.093343 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tkjv5" event={"ID":"68687a35-da55-497b-a7fa-0fe15b039d15","Type":"ContainerDied","Data":"b897c5f7a243e9735b79d27b9603072ef564e1639c66126a433f9b5e3038cbe8"} Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.093397 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b897c5f7a243e9735b79d27b9603072ef564e1639c66126a433f9b5e3038cbe8" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.093358 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tkjv5" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.096033 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hbb7f" event={"ID":"f6f3e489-f6ce-4126-adcb-764afd7e7159","Type":"ContainerStarted","Data":"3fbf9054afc050643754272d95d1919a8e11528c3ffdc041078fe5861e271a3d"} Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.097652 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-gjbv9" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.097653 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-gjbv9" event={"ID":"e29f134c-e329-4296-b491-101f71a255a0","Type":"ContainerDied","Data":"a8b644360a5cb31cde46e3f0c7798e85831e1f0496183f4235392cf9ab970f38"} Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.097724 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8b644360a5cb31cde46e3f0c7798e85831e1f0496183f4235392cf9ab970f38" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.103746 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"2aaa2a11-2896-4690-819b-61c6069ea6db","Type":"ContainerStarted","Data":"98934f68098f3a8ba412e497cebff7c26ee063396a235c4f79453de99331a4e4"} Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.105498 5035 generic.go:334] "Generic (PLEG): container finished" podID="97170441-9666-4bb1-856a-a229bca4810d" containerID="98958a9bc11d70d8847c9327d8c86fa88c1fcc75899eae7980e734aca897beb8" exitCode=0 Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.105553 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts-config-h96d9" event={"ID":"97170441-9666-4bb1-856a-a229bca4810d","Type":"ContainerDied","Data":"98958a9bc11d70d8847c9327d8c86fa88c1fcc75899eae7980e734aca897beb8"} Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.105584 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts-config-h96d9" event={"ID":"97170441-9666-4bb1-856a-a229bca4810d","Type":"ContainerStarted","Data":"271aab0182bfdf179fe5871cf918a7684461a7c9ed202f1a805920fe415510d5"} Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.156760 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.728660203 podStartE2EDuration="30.156740534s" podCreationTimestamp="2025-10-02 09:43:49 +0000 UTC" firstStartedPulling="2025-10-02 09:44:07.157675715 +0000 UTC m=+1012.514019740" lastFinishedPulling="2025-10-02 09:44:16.585756046 +0000 UTC m=+1021.942100071" observedRunningTime="2025-10-02 09:44:19.147776163 +0000 UTC m=+1024.504120188" watchObservedRunningTime="2025-10-02 09:44:19.156740534 +0000 UTC m=+1024.513084559" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.456512 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-2rxp7"] Oct 02 09:44:19 crc kubenswrapper[5035]: E1002 09:44:19.457065 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e29f134c-e329-4296-b491-101f71a255a0" containerName="mariadb-database-create" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.457077 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e29f134c-e329-4296-b491-101f71a255a0" containerName="mariadb-database-create" Oct 02 09:44:19 crc kubenswrapper[5035]: E1002 09:44:19.457117 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68687a35-da55-497b-a7fa-0fe15b039d15" containerName="mariadb-database-create" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.457123 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="68687a35-da55-497b-a7fa-0fe15b039d15" containerName="mariadb-database-create" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.457287 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e29f134c-e329-4296-b491-101f71a255a0" containerName="mariadb-database-create" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.457306 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="68687a35-da55-497b-a7fa-0fe15b039d15" containerName="mariadb-database-create" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.459949 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.465032 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.471410 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-2rxp7"] Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.551898 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjcss\" (UniqueName: \"kubernetes.io/projected/9f2fd41d-10e9-4f68-bf8e-8a482b827690-kube-api-access-bjcss\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.551946 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.551986 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.552011 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.552027 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-config\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.552085 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.653462 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.653571 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjcss\" (UniqueName: \"kubernetes.io/projected/9f2fd41d-10e9-4f68-bf8e-8a482b827690-kube-api-access-bjcss\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.653608 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.653655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.653686 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-config\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.653707 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.654739 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.656523 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.656861 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.657389 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-config\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.657870 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.709039 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjcss\" (UniqueName: \"kubernetes.io/projected/9f2fd41d-10e9-4f68-bf8e-8a482b827690-kube-api-access-bjcss\") pod \"dnsmasq-dns-77585f5f8c-2rxp7\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:19 crc kubenswrapper[5035]: I1002 09:44:19.783105 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.287692 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-2rxp7"] Oct 02 09:44:20 crc kubenswrapper[5035]: W1002 09:44:20.303950 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9f2fd41d_10e9_4f68_bf8e_8a482b827690.slice/crio-33e38d1624d4780eb8750c0483734da1ccaaaad53bdb9694e78aff9ea4c879b0 WatchSource:0}: Error finding container 33e38d1624d4780eb8750c0483734da1ccaaaad53bdb9694e78aff9ea4c879b0: Status 404 returned error can't find the container with id 33e38d1624d4780eb8750c0483734da1ccaaaad53bdb9694e78aff9ea4c879b0 Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.509006 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591440 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-additional-scripts\") pod \"97170441-9666-4bb1-856a-a229bca4810d\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591515 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-scripts\") pod \"97170441-9666-4bb1-856a-a229bca4810d\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591691 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run-ovn\") pod \"97170441-9666-4bb1-856a-a229bca4810d\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591723 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhfcc\" (UniqueName: \"kubernetes.io/projected/97170441-9666-4bb1-856a-a229bca4810d-kube-api-access-dhfcc\") pod \"97170441-9666-4bb1-856a-a229bca4810d\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591771 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-log-ovn\") pod \"97170441-9666-4bb1-856a-a229bca4810d\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591764 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "97170441-9666-4bb1-856a-a229bca4810d" (UID: "97170441-9666-4bb1-856a-a229bca4810d"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591857 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "97170441-9666-4bb1-856a-a229bca4810d" (UID: "97170441-9666-4bb1-856a-a229bca4810d"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591857 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run\") pod \"97170441-9666-4bb1-856a-a229bca4810d\" (UID: \"97170441-9666-4bb1-856a-a229bca4810d\") " Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.591897 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run" (OuterVolumeSpecName: "var-run") pod "97170441-9666-4bb1-856a-a229bca4810d" (UID: "97170441-9666-4bb1-856a-a229bca4810d"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.592831 5035 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.592845 5035 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.592856 5035 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/97170441-9666-4bb1-856a-a229bca4810d-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.592923 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-scripts" (OuterVolumeSpecName: "scripts") pod "97170441-9666-4bb1-856a-a229bca4810d" (UID: "97170441-9666-4bb1-856a-a229bca4810d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.593120 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "97170441-9666-4bb1-856a-a229bca4810d" (UID: "97170441-9666-4bb1-856a-a229bca4810d"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.596270 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97170441-9666-4bb1-856a-a229bca4810d-kube-api-access-dhfcc" (OuterVolumeSpecName: "kube-api-access-dhfcc") pod "97170441-9666-4bb1-856a-a229bca4810d" (UID: "97170441-9666-4bb1-856a-a229bca4810d"). InnerVolumeSpecName "kube-api-access-dhfcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.694073 5035 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.694124 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/97170441-9666-4bb1-856a-a229bca4810d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:20 crc kubenswrapper[5035]: I1002 09:44:20.694133 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhfcc\" (UniqueName: \"kubernetes.io/projected/97170441-9666-4bb1-856a-a229bca4810d-kube-api-access-dhfcc\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.139681 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jvdts-config-h96d9" event={"ID":"97170441-9666-4bb1-856a-a229bca4810d","Type":"ContainerDied","Data":"271aab0182bfdf179fe5871cf918a7684461a7c9ed202f1a805920fe415510d5"} Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.140029 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="271aab0182bfdf179fe5871cf918a7684461a7c9ed202f1a805920fe415510d5" Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.139724 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jvdts-config-h96d9" Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.142452 5035 generic.go:334] "Generic (PLEG): container finished" podID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerID="4ab336006445281cc0f01128884b576e880092ba1dc49158b120b76b24913468" exitCode=0 Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.142503 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" event={"ID":"9f2fd41d-10e9-4f68-bf8e-8a482b827690","Type":"ContainerDied","Data":"4ab336006445281cc0f01128884b576e880092ba1dc49158b120b76b24913468"} Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.142571 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" event={"ID":"9f2fd41d-10e9-4f68-bf8e-8a482b827690","Type":"ContainerStarted","Data":"33e38d1624d4780eb8750c0483734da1ccaaaad53bdb9694e78aff9ea4c879b0"} Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.605575 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jvdts-config-h96d9"] Oct 02 09:44:21 crc kubenswrapper[5035]: I1002 09:44:21.614179 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jvdts-config-h96d9"] Oct 02 09:44:22 crc kubenswrapper[5035]: I1002 09:44:22.175366 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97170441-9666-4bb1-856a-a229bca4810d" path="/var/lib/kubelet/pods/97170441-9666-4bb1-856a-a229bca4810d/volumes" Oct 02 09:44:23 crc kubenswrapper[5035]: I1002 09:44:23.809444 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.173014 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" event={"ID":"9f2fd41d-10e9-4f68-bf8e-8a482b827690","Type":"ContainerStarted","Data":"48f226ce96c69da3e41b507332f66fe23ed15441d0c4ac7aa86053a91d748990"} Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.173073 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.190447 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" podStartSLOduration=5.190427628 podStartE2EDuration="5.190427628s" podCreationTimestamp="2025-10-02 09:44:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:24.189574003 +0000 UTC m=+1029.545918038" watchObservedRunningTime="2025-10-02 09:44:24.190427628 +0000 UTC m=+1029.546771653" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.477951 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8a80-account-create-666mg"] Oct 02 09:44:24 crc kubenswrapper[5035]: E1002 09:44:24.478642 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97170441-9666-4bb1-856a-a229bca4810d" containerName="ovn-config" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.478665 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="97170441-9666-4bb1-856a-a229bca4810d" containerName="ovn-config" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.478891 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="97170441-9666-4bb1-856a-a229bca4810d" containerName="ovn-config" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.479471 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8a80-account-create-666mg" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.482669 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.487145 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8a80-account-create-666mg"] Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.574942 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4fxz\" (UniqueName: \"kubernetes.io/projected/37719aef-3ac7-486b-a59b-4acd555afcc1-kube-api-access-k4fxz\") pod \"cinder-8a80-account-create-666mg\" (UID: \"37719aef-3ac7-486b-a59b-4acd555afcc1\") " pod="openstack/cinder-8a80-account-create-666mg" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.678650 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4fxz\" (UniqueName: \"kubernetes.io/projected/37719aef-3ac7-486b-a59b-4acd555afcc1-kube-api-access-k4fxz\") pod \"cinder-8a80-account-create-666mg\" (UID: \"37719aef-3ac7-486b-a59b-4acd555afcc1\") " pod="openstack/cinder-8a80-account-create-666mg" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.682005 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6096-account-create-vxgpc"] Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.683269 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6096-account-create-vxgpc" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.685281 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.699809 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6096-account-create-vxgpc"] Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.705452 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4fxz\" (UniqueName: \"kubernetes.io/projected/37719aef-3ac7-486b-a59b-4acd555afcc1-kube-api-access-k4fxz\") pod \"cinder-8a80-account-create-666mg\" (UID: \"37719aef-3ac7-486b-a59b-4acd555afcc1\") " pod="openstack/cinder-8a80-account-create-666mg" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.780524 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qdvb\" (UniqueName: \"kubernetes.io/projected/50486824-cd3b-489e-986f-d5107331b8b1-kube-api-access-4qdvb\") pod \"neutron-6096-account-create-vxgpc\" (UID: \"50486824-cd3b-489e-986f-d5107331b8b1\") " pod="openstack/neutron-6096-account-create-vxgpc" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.810113 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8a80-account-create-666mg" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.881797 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qdvb\" (UniqueName: \"kubernetes.io/projected/50486824-cd3b-489e-986f-d5107331b8b1-kube-api-access-4qdvb\") pod \"neutron-6096-account-create-vxgpc\" (UID: \"50486824-cd3b-489e-986f-d5107331b8b1\") " pod="openstack/neutron-6096-account-create-vxgpc" Oct 02 09:44:24 crc kubenswrapper[5035]: I1002 09:44:24.898705 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qdvb\" (UniqueName: \"kubernetes.io/projected/50486824-cd3b-489e-986f-d5107331b8b1-kube-api-access-4qdvb\") pod \"neutron-6096-account-create-vxgpc\" (UID: \"50486824-cd3b-489e-986f-d5107331b8b1\") " pod="openstack/neutron-6096-account-create-vxgpc" Oct 02 09:44:25 crc kubenswrapper[5035]: I1002 09:44:25.010635 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6096-account-create-vxgpc" Oct 02 09:44:29 crc kubenswrapper[5035]: I1002 09:44:29.785941 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:44:29 crc kubenswrapper[5035]: I1002 09:44:29.838764 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-xcb6x"] Oct 02 09:44:29 crc kubenswrapper[5035]: I1002 09:44:29.839024 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-xcb6x" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerName="dnsmasq-dns" containerID="cri-o://ac79e8484380468920f1feaed79f4abc5c149f7365a87f83e5ef4a6dea908b60" gracePeriod=10 Oct 02 09:44:29 crc kubenswrapper[5035]: I1002 09:44:29.885111 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-xcb6x" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.235227 5035 generic.go:334] "Generic (PLEG): container finished" podID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerID="ac79e8484380468920f1feaed79f4abc5c149f7365a87f83e5ef4a6dea908b60" exitCode=0 Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.235342 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-xcb6x" event={"ID":"30d3793a-815a-4e7b-b3ec-6300de91a736","Type":"ContainerDied","Data":"ac79e8484380468920f1feaed79f4abc5c149f7365a87f83e5ef4a6dea908b60"} Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.294823 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.393885 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fstp9\" (UniqueName: \"kubernetes.io/projected/30d3793a-815a-4e7b-b3ec-6300de91a736-kube-api-access-fstp9\") pod \"30d3793a-815a-4e7b-b3ec-6300de91a736\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.393991 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-dns-svc\") pod \"30d3793a-815a-4e7b-b3ec-6300de91a736\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.394021 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-sb\") pod \"30d3793a-815a-4e7b-b3ec-6300de91a736\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.394733 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-config\") pod \"30d3793a-815a-4e7b-b3ec-6300de91a736\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.394795 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-nb\") pod \"30d3793a-815a-4e7b-b3ec-6300de91a736\" (UID: \"30d3793a-815a-4e7b-b3ec-6300de91a736\") " Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.399693 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d3793a-815a-4e7b-b3ec-6300de91a736-kube-api-access-fstp9" (OuterVolumeSpecName: "kube-api-access-fstp9") pod "30d3793a-815a-4e7b-b3ec-6300de91a736" (UID: "30d3793a-815a-4e7b-b3ec-6300de91a736"). InnerVolumeSpecName "kube-api-access-fstp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.442679 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "30d3793a-815a-4e7b-b3ec-6300de91a736" (UID: "30d3793a-815a-4e7b-b3ec-6300de91a736"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.442795 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-config" (OuterVolumeSpecName: "config") pod "30d3793a-815a-4e7b-b3ec-6300de91a736" (UID: "30d3793a-815a-4e7b-b3ec-6300de91a736"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.448275 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "30d3793a-815a-4e7b-b3ec-6300de91a736" (UID: "30d3793a-815a-4e7b-b3ec-6300de91a736"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.449433 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "30d3793a-815a-4e7b-b3ec-6300de91a736" (UID: "30d3793a-815a-4e7b-b3ec-6300de91a736"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.496742 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.496790 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.496805 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fstp9\" (UniqueName: \"kubernetes.io/projected/30d3793a-815a-4e7b-b3ec-6300de91a736-kube-api-access-fstp9\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.496817 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.496827 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/30d3793a-815a-4e7b-b3ec-6300de91a736-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.563573 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6096-account-create-vxgpc"] Oct 02 09:44:31 crc kubenswrapper[5035]: W1002 09:44:31.575219 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50486824_cd3b_489e_986f_d5107331b8b1.slice/crio-e7172d99d84b906cbc00bcf86a6a7a42e86eb1ddf07e917c17d689fb3690250a WatchSource:0}: Error finding container e7172d99d84b906cbc00bcf86a6a7a42e86eb1ddf07e917c17d689fb3690250a: Status 404 returned error can't find the container with id e7172d99d84b906cbc00bcf86a6a7a42e86eb1ddf07e917c17d689fb3690250a Oct 02 09:44:31 crc kubenswrapper[5035]: I1002 09:44:31.628163 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8a80-account-create-666mg"] Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.246691 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-xcb6x" event={"ID":"30d3793a-815a-4e7b-b3ec-6300de91a736","Type":"ContainerDied","Data":"6d03e94f44cd4114a77fe1b39792fe7c541f259ffa07fbffa739ac0c8dd1a7f2"} Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.247055 5035 scope.go:117] "RemoveContainer" containerID="ac79e8484380468920f1feaed79f4abc5c149f7365a87f83e5ef4a6dea908b60" Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.246754 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-xcb6x" Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.249442 5035 generic.go:334] "Generic (PLEG): container finished" podID="37719aef-3ac7-486b-a59b-4acd555afcc1" containerID="47ce2159e127d801ff299e7bf0640995b0ebb35ccd9d56e89cf3b1c4476ba02c" exitCode=0 Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.249526 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8a80-account-create-666mg" event={"ID":"37719aef-3ac7-486b-a59b-4acd555afcc1","Type":"ContainerDied","Data":"47ce2159e127d801ff299e7bf0640995b0ebb35ccd9d56e89cf3b1c4476ba02c"} Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.249573 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8a80-account-create-666mg" event={"ID":"37719aef-3ac7-486b-a59b-4acd555afcc1","Type":"ContainerStarted","Data":"875840a5fa49dc7e885f7267a9b43eb0b7df296850055eaf79fa109692a47e09"} Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.252627 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hbb7f" event={"ID":"f6f3e489-f6ce-4126-adcb-764afd7e7159","Type":"ContainerStarted","Data":"fd0b46b7da0368c1ededacc7b6bbcccc49d456086e7701378108a883922d0c96"} Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.255934 5035 generic.go:334] "Generic (PLEG): container finished" podID="50486824-cd3b-489e-986f-d5107331b8b1" containerID="3eb3fea218d76ef05e18df0c8a55add715a9fd458d3194af7aaad8e5c2bc5925" exitCode=0 Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.256113 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6096-account-create-vxgpc" event={"ID":"50486824-cd3b-489e-986f-d5107331b8b1","Type":"ContainerDied","Data":"3eb3fea218d76ef05e18df0c8a55add715a9fd458d3194af7aaad8e5c2bc5925"} Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.256141 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6096-account-create-vxgpc" event={"ID":"50486824-cd3b-489e-986f-d5107331b8b1","Type":"ContainerStarted","Data":"e7172d99d84b906cbc00bcf86a6a7a42e86eb1ddf07e917c17d689fb3690250a"} Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.266228 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhwgl" event={"ID":"5bf48931-dee2-4f51-a3dc-d69a4604064b","Type":"ContainerStarted","Data":"d2218c6995a9867a45454a9cf044c0a69107d1783052cf86a94dcb9b1227b334"} Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.268147 5035 scope.go:117] "RemoveContainer" containerID="401a1e90b66d3b939e272acee7a4a073be94d571167a055f65910165f45b4664" Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.281983 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-hbb7f" podStartSLOduration=3.084851425 podStartE2EDuration="15.281961721s" podCreationTimestamp="2025-10-02 09:44:17 +0000 UTC" firstStartedPulling="2025-10-02 09:44:18.905640511 +0000 UTC m=+1024.261984536" lastFinishedPulling="2025-10-02 09:44:31.102750807 +0000 UTC m=+1036.459094832" observedRunningTime="2025-10-02 09:44:32.27506861 +0000 UTC m=+1037.631412645" watchObservedRunningTime="2025-10-02 09:44:32.281961721 +0000 UTC m=+1037.638305746" Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.302053 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-xcb6x"] Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.310253 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-xcb6x"] Oct 02 09:44:32 crc kubenswrapper[5035]: I1002 09:44:32.376162 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-zhwgl" podStartSLOduration=9.959528898 podStartE2EDuration="15.376133884s" podCreationTimestamp="2025-10-02 09:44:17 +0000 UTC" firstStartedPulling="2025-10-02 09:44:18.734779105 +0000 UTC m=+1024.091123130" lastFinishedPulling="2025-10-02 09:44:24.151384091 +0000 UTC m=+1029.507728116" observedRunningTime="2025-10-02 09:44:32.37361083 +0000 UTC m=+1037.729954875" watchObservedRunningTime="2025-10-02 09:44:32.376133884 +0000 UTC m=+1037.732477909" Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.705667 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6096-account-create-vxgpc" Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.711255 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8a80-account-create-666mg" Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.812776 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.836088 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4fxz\" (UniqueName: \"kubernetes.io/projected/37719aef-3ac7-486b-a59b-4acd555afcc1-kube-api-access-k4fxz\") pod \"37719aef-3ac7-486b-a59b-4acd555afcc1\" (UID: \"37719aef-3ac7-486b-a59b-4acd555afcc1\") " Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.836249 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qdvb\" (UniqueName: \"kubernetes.io/projected/50486824-cd3b-489e-986f-d5107331b8b1-kube-api-access-4qdvb\") pod \"50486824-cd3b-489e-986f-d5107331b8b1\" (UID: \"50486824-cd3b-489e-986f-d5107331b8b1\") " Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.842396 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50486824-cd3b-489e-986f-d5107331b8b1-kube-api-access-4qdvb" (OuterVolumeSpecName: "kube-api-access-4qdvb") pod "50486824-cd3b-489e-986f-d5107331b8b1" (UID: "50486824-cd3b-489e-986f-d5107331b8b1"). InnerVolumeSpecName "kube-api-access-4qdvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.843612 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37719aef-3ac7-486b-a59b-4acd555afcc1-kube-api-access-k4fxz" (OuterVolumeSpecName: "kube-api-access-k4fxz") pod "37719aef-3ac7-486b-a59b-4acd555afcc1" (UID: "37719aef-3ac7-486b-a59b-4acd555afcc1"). InnerVolumeSpecName "kube-api-access-k4fxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.938931 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4fxz\" (UniqueName: \"kubernetes.io/projected/37719aef-3ac7-486b-a59b-4acd555afcc1-kube-api-access-k4fxz\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:33 crc kubenswrapper[5035]: I1002 09:44:33.938965 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qdvb\" (UniqueName: \"kubernetes.io/projected/50486824-cd3b-489e-986f-d5107331b8b1-kube-api-access-4qdvb\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.175468 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" path="/var/lib/kubelet/pods/30d3793a-815a-4e7b-b3ec-6300de91a736/volumes" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.286453 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8a80-account-create-666mg" event={"ID":"37719aef-3ac7-486b-a59b-4acd555afcc1","Type":"ContainerDied","Data":"875840a5fa49dc7e885f7267a9b43eb0b7df296850055eaf79fa109692a47e09"} Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.286498 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="875840a5fa49dc7e885f7267a9b43eb0b7df296850055eaf79fa109692a47e09" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.286567 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8a80-account-create-666mg" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.288828 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6096-account-create-vxgpc" event={"ID":"50486824-cd3b-489e-986f-d5107331b8b1","Type":"ContainerDied","Data":"e7172d99d84b906cbc00bcf86a6a7a42e86eb1ddf07e917c17d689fb3690250a"} Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.288855 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7172d99d84b906cbc00bcf86a6a7a42e86eb1ddf07e917c17d689fb3690250a" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.288889 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6096-account-create-vxgpc" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.501292 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-7c81-account-create-nk5lf"] Oct 02 09:44:34 crc kubenswrapper[5035]: E1002 09:44:34.501763 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerName="dnsmasq-dns" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.501781 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerName="dnsmasq-dns" Oct 02 09:44:34 crc kubenswrapper[5035]: E1002 09:44:34.501795 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37719aef-3ac7-486b-a59b-4acd555afcc1" containerName="mariadb-account-create" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.501809 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="37719aef-3ac7-486b-a59b-4acd555afcc1" containerName="mariadb-account-create" Oct 02 09:44:34 crc kubenswrapper[5035]: E1002 09:44:34.501817 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50486824-cd3b-489e-986f-d5107331b8b1" containerName="mariadb-account-create" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.501823 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="50486824-cd3b-489e-986f-d5107331b8b1" containerName="mariadb-account-create" Oct 02 09:44:34 crc kubenswrapper[5035]: E1002 09:44:34.501838 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerName="init" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.501846 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerName="init" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.502013 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d3793a-815a-4e7b-b3ec-6300de91a736" containerName="dnsmasq-dns" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.502036 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="50486824-cd3b-489e-986f-d5107331b8b1" containerName="mariadb-account-create" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.502048 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="37719aef-3ac7-486b-a59b-4acd555afcc1" containerName="mariadb-account-create" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.502646 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7c81-account-create-nk5lf" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.504838 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.507768 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7c81-account-create-nk5lf"] Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.547199 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6mlr\" (UniqueName: \"kubernetes.io/projected/0155abf0-b42c-4a52-adf4-cef5dfc9c1b6-kube-api-access-x6mlr\") pod \"barbican-7c81-account-create-nk5lf\" (UID: \"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6\") " pod="openstack/barbican-7c81-account-create-nk5lf" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.649363 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6mlr\" (UniqueName: \"kubernetes.io/projected/0155abf0-b42c-4a52-adf4-cef5dfc9c1b6-kube-api-access-x6mlr\") pod \"barbican-7c81-account-create-nk5lf\" (UID: \"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6\") " pod="openstack/barbican-7c81-account-create-nk5lf" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.671410 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6mlr\" (UniqueName: \"kubernetes.io/projected/0155abf0-b42c-4a52-adf4-cef5dfc9c1b6-kube-api-access-x6mlr\") pod \"barbican-7c81-account-create-nk5lf\" (UID: \"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6\") " pod="openstack/barbican-7c81-account-create-nk5lf" Oct 02 09:44:34 crc kubenswrapper[5035]: I1002 09:44:34.821387 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7c81-account-create-nk5lf" Oct 02 09:44:35 crc kubenswrapper[5035]: I1002 09:44:35.253599 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-7c81-account-create-nk5lf"] Oct 02 09:44:35 crc kubenswrapper[5035]: W1002 09:44:35.260434 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0155abf0_b42c_4a52_adf4_cef5dfc9c1b6.slice/crio-3d5c9af4571f00dfeb73a0dda5d66ae842e96d74427ddf0d5b0584aecfdb012a WatchSource:0}: Error finding container 3d5c9af4571f00dfeb73a0dda5d66ae842e96d74427ddf0d5b0584aecfdb012a: Status 404 returned error can't find the container with id 3d5c9af4571f00dfeb73a0dda5d66ae842e96d74427ddf0d5b0584aecfdb012a Oct 02 09:44:35 crc kubenswrapper[5035]: I1002 09:44:35.303488 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7c81-account-create-nk5lf" event={"ID":"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6","Type":"ContainerStarted","Data":"3d5c9af4571f00dfeb73a0dda5d66ae842e96d74427ddf0d5b0584aecfdb012a"} Oct 02 09:44:36 crc kubenswrapper[5035]: I1002 09:44:36.313901 5035 generic.go:334] "Generic (PLEG): container finished" podID="0155abf0-b42c-4a52-adf4-cef5dfc9c1b6" containerID="2db48747e2cf060ad12fdbfbc2f7b4c224d81ea2f5170ffee9e1d4692705084f" exitCode=0 Oct 02 09:44:36 crc kubenswrapper[5035]: I1002 09:44:36.314041 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7c81-account-create-nk5lf" event={"ID":"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6","Type":"ContainerDied","Data":"2db48747e2cf060ad12fdbfbc2f7b4c224d81ea2f5170ffee9e1d4692705084f"} Oct 02 09:44:37 crc kubenswrapper[5035]: I1002 09:44:37.325738 5035 generic.go:334] "Generic (PLEG): container finished" podID="5bf48931-dee2-4f51-a3dc-d69a4604064b" containerID="d2218c6995a9867a45454a9cf044c0a69107d1783052cf86a94dcb9b1227b334" exitCode=0 Oct 02 09:44:37 crc kubenswrapper[5035]: I1002 09:44:37.325940 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhwgl" event={"ID":"5bf48931-dee2-4f51-a3dc-d69a4604064b","Type":"ContainerDied","Data":"d2218c6995a9867a45454a9cf044c0a69107d1783052cf86a94dcb9b1227b334"} Oct 02 09:44:37 crc kubenswrapper[5035]: I1002 09:44:37.614887 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7c81-account-create-nk5lf" Oct 02 09:44:37 crc kubenswrapper[5035]: I1002 09:44:37.696298 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6mlr\" (UniqueName: \"kubernetes.io/projected/0155abf0-b42c-4a52-adf4-cef5dfc9c1b6-kube-api-access-x6mlr\") pod \"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6\" (UID: \"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6\") " Oct 02 09:44:37 crc kubenswrapper[5035]: I1002 09:44:37.701307 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0155abf0-b42c-4a52-adf4-cef5dfc9c1b6-kube-api-access-x6mlr" (OuterVolumeSpecName: "kube-api-access-x6mlr") pod "0155abf0-b42c-4a52-adf4-cef5dfc9c1b6" (UID: "0155abf0-b42c-4a52-adf4-cef5dfc9c1b6"). InnerVolumeSpecName "kube-api-access-x6mlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:37 crc kubenswrapper[5035]: I1002 09:44:37.798570 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6mlr\" (UniqueName: \"kubernetes.io/projected/0155abf0-b42c-4a52-adf4-cef5dfc9c1b6-kube-api-access-x6mlr\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.339122 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-7c81-account-create-nk5lf" event={"ID":"0155abf0-b42c-4a52-adf4-cef5dfc9c1b6","Type":"ContainerDied","Data":"3d5c9af4571f00dfeb73a0dda5d66ae842e96d74427ddf0d5b0584aecfdb012a"} Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.339174 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d5c9af4571f00dfeb73a0dda5d66ae842e96d74427ddf0d5b0584aecfdb012a" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.339148 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-7c81-account-create-nk5lf" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.597294 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.714758 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgxqt\" (UniqueName: \"kubernetes.io/projected/5bf48931-dee2-4f51-a3dc-d69a4604064b-kube-api-access-cgxqt\") pod \"5bf48931-dee2-4f51-a3dc-d69a4604064b\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.714853 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-config-data\") pod \"5bf48931-dee2-4f51-a3dc-d69a4604064b\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.714887 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-combined-ca-bundle\") pod \"5bf48931-dee2-4f51-a3dc-d69a4604064b\" (UID: \"5bf48931-dee2-4f51-a3dc-d69a4604064b\") " Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.733796 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bf48931-dee2-4f51-a3dc-d69a4604064b-kube-api-access-cgxqt" (OuterVolumeSpecName: "kube-api-access-cgxqt") pod "5bf48931-dee2-4f51-a3dc-d69a4604064b" (UID: "5bf48931-dee2-4f51-a3dc-d69a4604064b"). InnerVolumeSpecName "kube-api-access-cgxqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.745803 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bf48931-dee2-4f51-a3dc-d69a4604064b" (UID: "5bf48931-dee2-4f51-a3dc-d69a4604064b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.760494 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-config-data" (OuterVolumeSpecName: "config-data") pod "5bf48931-dee2-4f51-a3dc-d69a4604064b" (UID: "5bf48931-dee2-4f51-a3dc-d69a4604064b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.816419 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgxqt\" (UniqueName: \"kubernetes.io/projected/5bf48931-dee2-4f51-a3dc-d69a4604064b-kube-api-access-cgxqt\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.816470 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:38 crc kubenswrapper[5035]: I1002 09:44:38.816491 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bf48931-dee2-4f51-a3dc-d69a4604064b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.350390 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-zhwgl" event={"ID":"5bf48931-dee2-4f51-a3dc-d69a4604064b","Type":"ContainerDied","Data":"a0bb2b212c2c19aac312c2a2cc3b38b89b5d3ff4dc4f49d29b4c24e605b8d5e1"} Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.350816 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0bb2b212c2c19aac312c2a2cc3b38b89b5d3ff4dc4f49d29b4c24e605b8d5e1" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.350890 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-zhwgl" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.585378 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-h7npz"] Oct 02 09:44:39 crc kubenswrapper[5035]: E1002 09:44:39.585733 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf48931-dee2-4f51-a3dc-d69a4604064b" containerName="keystone-db-sync" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.585747 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf48931-dee2-4f51-a3dc-d69a4604064b" containerName="keystone-db-sync" Oct 02 09:44:39 crc kubenswrapper[5035]: E1002 09:44:39.585766 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0155abf0-b42c-4a52-adf4-cef5dfc9c1b6" containerName="mariadb-account-create" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.585801 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0155abf0-b42c-4a52-adf4-cef5dfc9c1b6" containerName="mariadb-account-create" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.586003 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf48931-dee2-4f51-a3dc-d69a4604064b" containerName="keystone-db-sync" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.586020 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0155abf0-b42c-4a52-adf4-cef5dfc9c1b6" containerName="mariadb-account-create" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.586825 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.599903 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-h7npz"] Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.632514 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.632764 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-svc\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.632841 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-kube-api-access-vnjlp\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.632874 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.632946 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-config\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.632996 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.641005 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5cvfn"] Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.642114 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.644165 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.644580 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.644948 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.648862 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fj9kx" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.662998 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5cvfn"] Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.734900 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-config\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.734947 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-fernet-keys\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.734968 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735006 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735031 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-credential-keys\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735075 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-scripts\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735095 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-combined-ca-bundle\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735120 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-config-data\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735142 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jczv9\" (UniqueName: \"kubernetes.io/projected/942a97bf-0aac-4dec-ba19-34e188adc514-kube-api-access-jczv9\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735162 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-svc\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735189 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-kube-api-access-vnjlp\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.735205 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.736190 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.736763 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-config\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.737332 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.737791 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-svc\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.741168 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.768434 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-kube-api-access-vnjlp\") pod \"dnsmasq-dns-55fff446b9-h7npz\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.837552 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-fernet-keys\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.837655 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-credential-keys\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.837713 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-scripts\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.837735 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-combined-ca-bundle\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.837767 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-config-data\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.837801 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jczv9\" (UniqueName: \"kubernetes.io/projected/942a97bf-0aac-4dec-ba19-34e188adc514-kube-api-access-jczv9\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.839140 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-create-mqjfl"] Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.857251 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-fernet-keys\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.867737 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-mqjfl" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.879326 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-config-data\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.879439 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-scripts\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.882278 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-credential-keys\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.882308 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-combined-ca-bundle\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.888485 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-mqjfl"] Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.889681 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jczv9\" (UniqueName: \"kubernetes.io/projected/942a97bf-0aac-4dec-ba19-34e188adc514-kube-api-access-jczv9\") pod \"keystone-bootstrap-5cvfn\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.912800 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.915782 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7t2tz"] Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.917220 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.918689 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.924885 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.931892 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8nccm" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.939561 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8blx\" (UniqueName: \"kubernetes.io/projected/aa5bbaa2-0524-45c3-9643-6bc946e45870-kube-api-access-w8blx\") pod \"ironic-db-create-mqjfl\" (UID: \"aa5bbaa2-0524-45c3-9643-6bc946e45870\") " pod="openstack/ironic-db-create-mqjfl" Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.951737 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7t2tz"] Oct 02 09:44:39 crc kubenswrapper[5035]: I1002 09:44:39.963942 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.045595 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-config-data\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.045959 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8blx\" (UniqueName: \"kubernetes.io/projected/aa5bbaa2-0524-45c3-9643-6bc946e45870-kube-api-access-w8blx\") pod \"ironic-db-create-mqjfl\" (UID: \"aa5bbaa2-0524-45c3-9643-6bc946e45870\") " pod="openstack/ironic-db-create-mqjfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.046014 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-combined-ca-bundle\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.046064 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-db-sync-config-data\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.046120 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpm2v\" (UniqueName: \"kubernetes.io/projected/76ce8ea1-cda3-4f68-9570-c6c02b56283f-kube-api-access-mpm2v\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.046145 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-scripts\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.046202 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ce8ea1-cda3-4f68-9570-c6c02b56283f-etc-machine-id\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.056608 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-h7npz"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.076791 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-qf6x5"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.095349 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8blx\" (UniqueName: \"kubernetes.io/projected/aa5bbaa2-0524-45c3-9643-6bc946e45870-kube-api-access-w8blx\") pod \"ironic-db-create-mqjfl\" (UID: \"aa5bbaa2-0524-45c3-9643-6bc946e45870\") " pod="openstack/ironic-db-create-mqjfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.097700 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.107053 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-phxhr" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.107355 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.107552 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-knpjm"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.108656 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.110458 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.111070 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.113401 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f7xgd" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.125281 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qf6x5"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.132664 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-knpjm"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.144449 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.146773 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.147699 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-config-data\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.147750 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-combined-ca-bundle\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.147785 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-config-data\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.147799 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9rvk\" (UniqueName: \"kubernetes.io/projected/830b460b-be63-4b42-a442-80d844ef6908-kube-api-access-f9rvk\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.147816 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136b4da6-5c57-4a29-af5a-afe32c5f552f-logs\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148090 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-combined-ca-bundle\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148130 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-db-sync-config-data\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148148 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpm2v\" (UniqueName: \"kubernetes.io/projected/76ce8ea1-cda3-4f68-9570-c6c02b56283f-kube-api-access-mpm2v\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148169 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2k7k\" (UniqueName: \"kubernetes.io/projected/136b4da6-5c57-4a29-af5a-afe32c5f552f-kube-api-access-t2k7k\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148202 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-scripts\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148235 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-scripts\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148274 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-db-sync-config-data\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148393 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ce8ea1-cda3-4f68-9570-c6c02b56283f-etc-machine-id\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148464 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-combined-ca-bundle\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.148850 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.152410 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ce8ea1-cda3-4f68-9570-c6c02b56283f-etc-machine-id\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.152901 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.153352 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.157390 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-db-sync-config-data\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.158910 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-scripts\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.161359 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-config-data\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.161610 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-combined-ca-bundle\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.175505 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpm2v\" (UniqueName: \"kubernetes.io/projected/76ce8ea1-cda3-4f68-9570-c6c02b56283f-kube-api-access-mpm2v\") pod \"cinder-db-sync-7t2tz\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.190452 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-h8tfl"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.191899 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-h8tfl"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.191987 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.195834 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-d6j5s"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.197831 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.200483 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-46h8t" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.200682 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.200881 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.203501 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d6j5s"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251313 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2k7k\" (UniqueName: \"kubernetes.io/projected/136b4da6-5c57-4a29-af5a-afe32c5f552f-kube-api-access-t2k7k\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251655 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-scripts\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251677 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-scripts\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251695 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251713 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251753 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-config-data\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251776 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-db-sync-config-data\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251836 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-log-httpd\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251883 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-combined-ca-bundle\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251911 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-combined-ca-bundle\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251940 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251961 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-config\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.251983 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r88rm\" (UniqueName: \"kubernetes.io/projected/3636a749-409c-4b58-a29f-8ca1568d2dbc-kube-api-access-r88rm\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252124 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252186 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7cd8\" (UniqueName: \"kubernetes.io/projected/47413f27-08ae-470d-8877-84b1fd25fd8a-kube-api-access-g7cd8\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252234 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-combined-ca-bundle\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252296 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-config-data\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252331 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9rvk\" (UniqueName: \"kubernetes.io/projected/830b460b-be63-4b42-a442-80d844ef6908-kube-api-access-f9rvk\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136b4da6-5c57-4a29-af5a-afe32c5f552f-logs\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252393 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252441 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95x2t\" (UniqueName: \"kubernetes.io/projected/fa2948a0-da02-4c95-bc2b-d9789564c9e7-kube-api-access-95x2t\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252493 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-config\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252517 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.252751 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-run-httpd\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.254488 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136b4da6-5c57-4a29-af5a-afe32c5f552f-logs\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.257902 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-scripts\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.258148 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-db-sync-config-data\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.258318 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-combined-ca-bundle\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.258704 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-config-data\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.259029 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-combined-ca-bundle\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.270504 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9rvk\" (UniqueName: \"kubernetes.io/projected/830b460b-be63-4b42-a442-80d844ef6908-kube-api-access-f9rvk\") pod \"barbican-db-sync-qf6x5\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.270686 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2k7k\" (UniqueName: \"kubernetes.io/projected/136b4da6-5c57-4a29-af5a-afe32c5f552f-kube-api-access-t2k7k\") pod \"placement-db-sync-knpjm\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.335676 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-mqjfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354446 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7cd8\" (UniqueName: \"kubernetes.io/projected/47413f27-08ae-470d-8877-84b1fd25fd8a-kube-api-access-g7cd8\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354507 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354562 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95x2t\" (UniqueName: \"kubernetes.io/projected/fa2948a0-da02-4c95-bc2b-d9789564c9e7-kube-api-access-95x2t\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354596 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-config\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354617 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354639 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-run-httpd\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354689 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-scripts\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354722 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354742 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354774 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-config-data\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354831 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-log-httpd\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354861 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-combined-ca-bundle\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354897 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-config\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354946 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r88rm\" (UniqueName: \"kubernetes.io/projected/3636a749-409c-4b58-a29f-8ca1568d2dbc-kube-api-access-r88rm\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.354967 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.356285 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.357014 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.358105 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-config\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.381144 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.381596 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-combined-ca-bundle\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.381975 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-log-httpd\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.382494 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-run-httpd\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.383054 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-config\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.384786 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.385734 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.385782 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.385962 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-scripts\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.389752 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-config-data\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.401666 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r88rm\" (UniqueName: \"kubernetes.io/projected/3636a749-409c-4b58-a29f-8ca1568d2dbc-kube-api-access-r88rm\") pod \"ceilometer-0\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.404325 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7cd8\" (UniqueName: \"kubernetes.io/projected/47413f27-08ae-470d-8877-84b1fd25fd8a-kube-api-access-g7cd8\") pod \"neutron-db-sync-d6j5s\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.407831 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95x2t\" (UniqueName: \"kubernetes.io/projected/fa2948a0-da02-4c95-bc2b-d9789564c9e7-kube-api-access-95x2t\") pod \"dnsmasq-dns-76fcf4b695-h8tfl\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.412516 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.465135 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:44:40 crc kubenswrapper[5035]: W1002 09:44:40.465757 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod942a97bf_0aac_4dec_ba19_34e188adc514.slice/crio-11e295024087e98211503e16609b7e1065a621ee8c83c8f988570a8e8a13a97e WatchSource:0}: Error finding container 11e295024087e98211503e16609b7e1065a621ee8c83c8f988570a8e8a13a97e: Status 404 returned error can't find the container with id 11e295024087e98211503e16609b7e1065a621ee8c83c8f988570a8e8a13a97e Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.476141 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5cvfn"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.493662 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-knpjm" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.516267 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.533846 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.547551 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.593432 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-h7npz"] Oct 02 09:44:40 crc kubenswrapper[5035]: I1002 09:44:40.906103 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-mqjfl"] Oct 02 09:44:40 crc kubenswrapper[5035]: W1002 09:44:40.913710 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa5bbaa2_0524_45c3_9643_6bc946e45870.slice/crio-3ca1402894441c9f8166a347882f85c43286688691f7f8661a8643c1665e8af6 WatchSource:0}: Error finding container 3ca1402894441c9f8166a347882f85c43286688691f7f8661a8643c1665e8af6: Status 404 returned error can't find the container with id 3ca1402894441c9f8166a347882f85c43286688691f7f8661a8643c1665e8af6 Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.056075 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7t2tz"] Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.124143 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-qf6x5"] Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.214039 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-knpjm"] Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.235851 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:44:41 crc kubenswrapper[5035]: W1002 09:44:41.242287 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3636a749_409c_4b58_a29f_8ca1568d2dbc.slice/crio-ff97859ea3e14f77f87fda3eb564364b6d16d9caef16589bb892ca19cc9e882a WatchSource:0}: Error finding container ff97859ea3e14f77f87fda3eb564364b6d16d9caef16589bb892ca19cc9e882a: Status 404 returned error can't find the container with id ff97859ea3e14f77f87fda3eb564364b6d16d9caef16589bb892ca19cc9e882a Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.440851 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerStarted","Data":"ff97859ea3e14f77f87fda3eb564364b6d16d9caef16589bb892ca19cc9e882a"} Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.443017 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-h7npz" event={"ID":"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b","Type":"ContainerStarted","Data":"834769a99f048112071e0e61a6b51c066ce0b883b3ef5ef04c06ae2d4b5bbfa8"} Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.445192 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7t2tz" event={"ID":"76ce8ea1-cda3-4f68-9570-c6c02b56283f","Type":"ContainerStarted","Data":"8ff5d586b5d8e464ecb7fadfe3abe942d5c27229df686a200838c87b458a6587"} Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.454021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5cvfn" event={"ID":"942a97bf-0aac-4dec-ba19-34e188adc514","Type":"ContainerStarted","Data":"11e295024087e98211503e16609b7e1065a621ee8c83c8f988570a8e8a13a97e"} Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.456505 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-d6j5s"] Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.460419 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-knpjm" event={"ID":"136b4da6-5c57-4a29-af5a-afe32c5f552f","Type":"ContainerStarted","Data":"e78bfb5af27eaa5a80ca68cedcb1d6b7e1447fc366c25bae0e2944ad4d67d7e8"} Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.464134 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-mqjfl" event={"ID":"aa5bbaa2-0524-45c3-9643-6bc946e45870","Type":"ContainerStarted","Data":"3ca1402894441c9f8166a347882f85c43286688691f7f8661a8643c1665e8af6"} Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.465146 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qf6x5" event={"ID":"830b460b-be63-4b42-a442-80d844ef6908","Type":"ContainerStarted","Data":"41e6d8f29f456dd758a6eefe2b97cd495b0d9fe972266a920c8379ff09ea006d"} Oct 02 09:44:41 crc kubenswrapper[5035]: W1002 09:44:41.467140 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47413f27_08ae_470d_8877_84b1fd25fd8a.slice/crio-3acf39cbd430ffbcf8dfc9049422d0dbbd5414a7654ee36150ec6c8dfeadf78b WatchSource:0}: Error finding container 3acf39cbd430ffbcf8dfc9049422d0dbbd5414a7654ee36150ec6c8dfeadf78b: Status 404 returned error can't find the container with id 3acf39cbd430ffbcf8dfc9049422d0dbbd5414a7654ee36150ec6c8dfeadf78b Oct 02 09:44:41 crc kubenswrapper[5035]: I1002 09:44:41.494261 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-h8tfl"] Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.018135 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.506697 5035 generic.go:334] "Generic (PLEG): container finished" podID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerID="24dcf757f3fc909e46ee7b98193be6bd260d929e7cf0b4f62540c65634108952" exitCode=0 Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.506786 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" event={"ID":"fa2948a0-da02-4c95-bc2b-d9789564c9e7","Type":"ContainerDied","Data":"24dcf757f3fc909e46ee7b98193be6bd260d929e7cf0b4f62540c65634108952"} Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.510271 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" event={"ID":"fa2948a0-da02-4c95-bc2b-d9789564c9e7","Type":"ContainerStarted","Data":"db294ffa2d3eca59e8f6c713f2ff3b19ca901ec6e664878ab82bb6672aa44ee8"} Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.514623 5035 generic.go:334] "Generic (PLEG): container finished" podID="aa5bbaa2-0524-45c3-9643-6bc946e45870" containerID="5c0780b6a2f1cfeefbb3baf3deb3f58cc4fbf0e9169e435c3d96f74f4e4393a9" exitCode=0 Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.514710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-mqjfl" event={"ID":"aa5bbaa2-0524-45c3-9643-6bc946e45870","Type":"ContainerDied","Data":"5c0780b6a2f1cfeefbb3baf3deb3f58cc4fbf0e9169e435c3d96f74f4e4393a9"} Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.544317 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d6j5s" event={"ID":"47413f27-08ae-470d-8877-84b1fd25fd8a","Type":"ContainerStarted","Data":"53ac0626fdd0ddba759654706b8203018a894f59965851289a2b8475b311e981"} Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.544363 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d6j5s" event={"ID":"47413f27-08ae-470d-8877-84b1fd25fd8a","Type":"ContainerStarted","Data":"3acf39cbd430ffbcf8dfc9049422d0dbbd5414a7654ee36150ec6c8dfeadf78b"} Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.564398 5035 generic.go:334] "Generic (PLEG): container finished" podID="0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" containerID="cee277bb187ab5f1b331405e63adf9eccb96c29ee6c02576c063fe73f71862ee" exitCode=0 Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.564500 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-h7npz" event={"ID":"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b","Type":"ContainerDied","Data":"cee277bb187ab5f1b331405e63adf9eccb96c29ee6c02576c063fe73f71862ee"} Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.569729 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5cvfn" event={"ID":"942a97bf-0aac-4dec-ba19-34e188adc514","Type":"ContainerStarted","Data":"3f51f48e5f9e423c6395355aebd9732b5fcc9b082f94afe2a777f3f5890a3cea"} Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.585475 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-d6j5s" podStartSLOduration=2.585051034 podStartE2EDuration="2.585051034s" podCreationTimestamp="2025-10-02 09:44:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:42.574811616 +0000 UTC m=+1047.931155651" watchObservedRunningTime="2025-10-02 09:44:42.585051034 +0000 UTC m=+1047.941395059" Oct 02 09:44:42 crc kubenswrapper[5035]: I1002 09:44:42.600105 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5cvfn" podStartSLOduration=3.600083372 podStartE2EDuration="3.600083372s" podCreationTimestamp="2025-10-02 09:44:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:42.593461769 +0000 UTC m=+1047.949805794" watchObservedRunningTime="2025-10-02 09:44:42.600083372 +0000 UTC m=+1047.956427407" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.017358 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.122909 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-kube-api-access-vnjlp\") pod \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.123041 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-swift-storage-0\") pod \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.123195 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-config\") pod \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.123222 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-nb\") pod \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.123344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-sb\") pod \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.123403 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-svc\") pod \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\" (UID: \"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b\") " Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.129310 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-kube-api-access-vnjlp" (OuterVolumeSpecName: "kube-api-access-vnjlp") pod "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" (UID: "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b"). InnerVolumeSpecName "kube-api-access-vnjlp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.152922 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" (UID: "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.152964 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" (UID: "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.158258 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-config" (OuterVolumeSpecName: "config") pod "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" (UID: "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.164292 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" (UID: "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.167154 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" (UID: "0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.224829 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.224863 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.224875 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.224884 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.224893 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnjlp\" (UniqueName: \"kubernetes.io/projected/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-kube-api-access-vnjlp\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.224902 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.604765 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-h7npz" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.605752 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-h7npz" event={"ID":"0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b","Type":"ContainerDied","Data":"834769a99f048112071e0e61a6b51c066ce0b883b3ef5ef04c06ae2d4b5bbfa8"} Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.605828 5035 scope.go:117] "RemoveContainer" containerID="cee277bb187ab5f1b331405e63adf9eccb96c29ee6c02576c063fe73f71862ee" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.618209 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" event={"ID":"fa2948a0-da02-4c95-bc2b-d9789564c9e7","Type":"ContainerStarted","Data":"116c59c19ce9de2fa774d4767abea7fe77a099dde4c885273caa655d9818a6fc"} Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.618901 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.644452 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" podStartSLOduration=3.644427299 podStartE2EDuration="3.644427299s" podCreationTimestamp="2025-10-02 09:44:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:44:43.637548359 +0000 UTC m=+1048.993892384" watchObservedRunningTime="2025-10-02 09:44:43.644427299 +0000 UTC m=+1049.000771324" Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.697757 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-h7npz"] Oct 02 09:44:43 crc kubenswrapper[5035]: I1002 09:44:43.710417 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-h7npz"] Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.019047 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-mqjfl" Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.146817 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8blx\" (UniqueName: \"kubernetes.io/projected/aa5bbaa2-0524-45c3-9643-6bc946e45870-kube-api-access-w8blx\") pod \"aa5bbaa2-0524-45c3-9643-6bc946e45870\" (UID: \"aa5bbaa2-0524-45c3-9643-6bc946e45870\") " Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.158648 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5bbaa2-0524-45c3-9643-6bc946e45870-kube-api-access-w8blx" (OuterVolumeSpecName: "kube-api-access-w8blx") pod "aa5bbaa2-0524-45c3-9643-6bc946e45870" (UID: "aa5bbaa2-0524-45c3-9643-6bc946e45870"). InnerVolumeSpecName "kube-api-access-w8blx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.173662 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" path="/var/lib/kubelet/pods/0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b/volumes" Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.249716 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8blx\" (UniqueName: \"kubernetes.io/projected/aa5bbaa2-0524-45c3-9643-6bc946e45870-kube-api-access-w8blx\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.637921 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-mqjfl" Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.638628 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-mqjfl" event={"ID":"aa5bbaa2-0524-45c3-9643-6bc946e45870","Type":"ContainerDied","Data":"3ca1402894441c9f8166a347882f85c43286688691f7f8661a8643c1665e8af6"} Oct 02 09:44:44 crc kubenswrapper[5035]: I1002 09:44:44.638655 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ca1402894441c9f8166a347882f85c43286688691f7f8661a8643c1665e8af6" Oct 02 09:44:46 crc kubenswrapper[5035]: I1002 09:44:46.665211 5035 generic.go:334] "Generic (PLEG): container finished" podID="f6f3e489-f6ce-4126-adcb-764afd7e7159" containerID="fd0b46b7da0368c1ededacc7b6bbcccc49d456086e7701378108a883922d0c96" exitCode=0 Oct 02 09:44:46 crc kubenswrapper[5035]: I1002 09:44:46.665328 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hbb7f" event={"ID":"f6f3e489-f6ce-4126-adcb-764afd7e7159","Type":"ContainerDied","Data":"fd0b46b7da0368c1ededacc7b6bbcccc49d456086e7701378108a883922d0c96"} Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.621400 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.691785 5035 generic.go:334] "Generic (PLEG): container finished" podID="942a97bf-0aac-4dec-ba19-34e188adc514" containerID="3f51f48e5f9e423c6395355aebd9732b5fcc9b082f94afe2a777f3f5890a3cea" exitCode=0 Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.691845 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5cvfn" event={"ID":"942a97bf-0aac-4dec-ba19-34e188adc514","Type":"ContainerDied","Data":"3f51f48e5f9e423c6395355aebd9732b5fcc9b082f94afe2a777f3f5890a3cea"} Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.695042 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-hbb7f" event={"ID":"f6f3e489-f6ce-4126-adcb-764afd7e7159","Type":"ContainerDied","Data":"3fbf9054afc050643754272d95d1919a8e11528c3ffdc041078fe5861e271a3d"} Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.695103 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fbf9054afc050643754272d95d1919a8e11528c3ffdc041078fe5861e271a3d" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.695106 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-hbb7f" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.751904 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-config-data\") pod \"f6f3e489-f6ce-4126-adcb-764afd7e7159\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.752004 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-combined-ca-bundle\") pod \"f6f3e489-f6ce-4126-adcb-764afd7e7159\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.752022 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfnds\" (UniqueName: \"kubernetes.io/projected/f6f3e489-f6ce-4126-adcb-764afd7e7159-kube-api-access-kfnds\") pod \"f6f3e489-f6ce-4126-adcb-764afd7e7159\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.752124 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-db-sync-config-data\") pod \"f6f3e489-f6ce-4126-adcb-764afd7e7159\" (UID: \"f6f3e489-f6ce-4126-adcb-764afd7e7159\") " Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.759785 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f6f3e489-f6ce-4126-adcb-764afd7e7159" (UID: "f6f3e489-f6ce-4126-adcb-764afd7e7159"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.762377 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f3e489-f6ce-4126-adcb-764afd7e7159-kube-api-access-kfnds" (OuterVolumeSpecName: "kube-api-access-kfnds") pod "f6f3e489-f6ce-4126-adcb-764afd7e7159" (UID: "f6f3e489-f6ce-4126-adcb-764afd7e7159"). InnerVolumeSpecName "kube-api-access-kfnds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.785317 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6f3e489-f6ce-4126-adcb-764afd7e7159" (UID: "f6f3e489-f6ce-4126-adcb-764afd7e7159"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.814704 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-config-data" (OuterVolumeSpecName: "config-data") pod "f6f3e489-f6ce-4126-adcb-764afd7e7159" (UID: "f6f3e489-f6ce-4126-adcb-764afd7e7159"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.854388 5035 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.854425 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.854435 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6f3e489-f6ce-4126-adcb-764afd7e7159-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:49 crc kubenswrapper[5035]: I1002 09:44:49.854443 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfnds\" (UniqueName: \"kubernetes.io/projected/f6f3e489-f6ce-4126-adcb-764afd7e7159-kube-api-access-kfnds\") on node \"crc\" DevicePath \"\"" Oct 02 09:44:50 crc kubenswrapper[5035]: I1002 09:44:50.536016 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:44:50 crc kubenswrapper[5035]: I1002 09:44:50.606216 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-2rxp7"] Oct 02 09:44:50 crc kubenswrapper[5035]: I1002 09:44:50.606628 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="dnsmasq-dns" containerID="cri-o://48f226ce96c69da3e41b507332f66fe23ed15441d0c4ac7aa86053a91d748990" gracePeriod=10 Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.101166 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-286pl"] Oct 02 09:44:51 crc kubenswrapper[5035]: E1002 09:44:51.101491 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f3e489-f6ce-4126-adcb-764afd7e7159" containerName="glance-db-sync" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.101503 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f3e489-f6ce-4126-adcb-764afd7e7159" containerName="glance-db-sync" Oct 02 09:44:51 crc kubenswrapper[5035]: E1002 09:44:51.101513 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5bbaa2-0524-45c3-9643-6bc946e45870" containerName="mariadb-database-create" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.101519 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5bbaa2-0524-45c3-9643-6bc946e45870" containerName="mariadb-database-create" Oct 02 09:44:51 crc kubenswrapper[5035]: E1002 09:44:51.101558 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" containerName="init" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.101564 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" containerName="init" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.101711 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5bbaa2-0524-45c3-9643-6bc946e45870" containerName="mariadb-database-create" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.101729 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f6e1734-0e8a-4c6f-ad86-61d3c6b6288b" containerName="init" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.101743 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f3e489-f6ce-4126-adcb-764afd7e7159" containerName="glance-db-sync" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.102557 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.139255 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-286pl"] Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.181357 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5q6g\" (UniqueName: \"kubernetes.io/projected/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-kube-api-access-z5q6g\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.181512 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.181564 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.181588 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-config\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.181691 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.181743 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.283554 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.283628 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5q6g\" (UniqueName: \"kubernetes.io/projected/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-kube-api-access-z5q6g\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.283821 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.283877 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.283907 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-config\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.283992 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.284981 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.285606 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.289114 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.289977 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.290634 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-config\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.315200 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5q6g\" (UniqueName: \"kubernetes.io/projected/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-kube-api-access-z5q6g\") pod \"dnsmasq-dns-8b5c85b87-286pl\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.427909 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.711272 5035 generic.go:334] "Generic (PLEG): container finished" podID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerID="48f226ce96c69da3e41b507332f66fe23ed15441d0c4ac7aa86053a91d748990" exitCode=0 Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.711314 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" event={"ID":"9f2fd41d-10e9-4f68-bf8e-8a482b827690","Type":"ContainerDied","Data":"48f226ce96c69da3e41b507332f66fe23ed15441d0c4ac7aa86053a91d748990"} Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.994659 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.996107 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.998495 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2l9tg" Oct 02 09:44:51 crc kubenswrapper[5035]: I1002 09:44:51.998824 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.004499 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.020094 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.097099 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.099323 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.099477 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.099764 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-logs\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.100008 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.100063 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt5vs\" (UniqueName: \"kubernetes.io/projected/ce7048e6-3e78-4655-b89a-9f010dcec708-kube-api-access-dt5vs\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.100150 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.202312 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.202371 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt5vs\" (UniqueName: \"kubernetes.io/projected/ce7048e6-3e78-4655-b89a-9f010dcec708-kube-api-access-dt5vs\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.202431 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.202840 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.203148 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.203419 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.203720 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.203854 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-logs\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.204251 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-logs\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.204981 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.208396 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-config-data\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.208958 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-scripts\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.216080 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.224062 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt5vs\" (UniqueName: \"kubernetes.io/projected/ce7048e6-3e78-4655-b89a-9f010dcec708-kube-api-access-dt5vs\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.229278 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.320611 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.343046 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.346504 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.349477 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.356475 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.407657 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.407716 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.407739 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.407764 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.407801 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzw7f\" (UniqueName: \"kubernetes.io/projected/6e598700-b625-46eb-9807-9f81dc5e7bc5-kube-api-access-nzw7f\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.407820 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.407997 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.509708 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzw7f\" (UniqueName: \"kubernetes.io/projected/6e598700-b625-46eb-9807-9f81dc5e7bc5-kube-api-access-nzw7f\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.509771 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.509893 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.509943 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.509972 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.509992 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.510021 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.510360 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.510623 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.510652 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-logs\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.514523 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.514618 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.516220 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.529440 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzw7f\" (UniqueName: \"kubernetes.io/projected/6e598700-b625-46eb-9807-9f81dc5e7bc5-kube-api-access-nzw7f\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.531849 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:44:52 crc kubenswrapper[5035]: I1002 09:44:52.673183 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:44:53 crc kubenswrapper[5035]: I1002 09:44:53.920094 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:44:53 crc kubenswrapper[5035]: I1002 09:44:53.978357 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:44:54 crc kubenswrapper[5035]: E1002 09:44:54.245685 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 02 09:44:54 crc kubenswrapper[5035]: E1002 09:44:54.245858 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t2k7k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-knpjm_openstack(136b4da6-5c57-4a29-af5a-afe32c5f552f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:44:54 crc kubenswrapper[5035]: E1002 09:44:54.247221 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-knpjm" podUID="136b4da6-5c57-4a29-af5a-afe32c5f552f" Oct 02 09:44:54 crc kubenswrapper[5035]: E1002 09:44:54.733692 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-knpjm" podUID="136b4da6-5c57-4a29-af5a-afe32c5f552f" Oct 02 09:44:54 crc kubenswrapper[5035]: I1002 09:44:54.784376 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Oct 02 09:44:59 crc kubenswrapper[5035]: I1002 09:44:59.784139 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Oct 02 09:44:59 crc kubenswrapper[5035]: I1002 09:44:59.998334 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-8c07-account-create-9t9l8"] Oct 02 09:44:59 crc kubenswrapper[5035]: I1002 09:44:59.999439 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8c07-account-create-9t9l8" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.002506 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-db-secret" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.008585 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-8c07-account-create-9t9l8"] Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.052118 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5ksm\" (UniqueName: \"kubernetes.io/projected/2887e3dc-2301-42c6-ac2d-725adefd5fcf-kube-api-access-w5ksm\") pod \"ironic-8c07-account-create-9t9l8\" (UID: \"2887e3dc-2301-42c6-ac2d-725adefd5fcf\") " pod="openstack/ironic-8c07-account-create-9t9l8" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.141066 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv"] Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.142412 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.144762 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.144814 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.155986 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5ksm\" (UniqueName: \"kubernetes.io/projected/2887e3dc-2301-42c6-ac2d-725adefd5fcf-kube-api-access-w5ksm\") pod \"ironic-8c07-account-create-9t9l8\" (UID: \"2887e3dc-2301-42c6-ac2d-725adefd5fcf\") " pod="openstack/ironic-8c07-account-create-9t9l8" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.179944 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5ksm\" (UniqueName: \"kubernetes.io/projected/2887e3dc-2301-42c6-ac2d-725adefd5fcf-kube-api-access-w5ksm\") pod \"ironic-8c07-account-create-9t9l8\" (UID: \"2887e3dc-2301-42c6-ac2d-725adefd5fcf\") " pod="openstack/ironic-8c07-account-create-9t9l8" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.198598 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv"] Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.258096 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhvhz\" (UniqueName: \"kubernetes.io/projected/44426e56-e34e-4df8-b0b0-68de44efe94e-kube-api-access-xhvhz\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.258508 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44426e56-e34e-4df8-b0b0-68de44efe94e-secret-volume\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.258658 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44426e56-e34e-4df8-b0b0-68de44efe94e-config-volume\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.327277 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8c07-account-create-9t9l8" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.360149 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44426e56-e34e-4df8-b0b0-68de44efe94e-secret-volume\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.360204 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44426e56-e34e-4df8-b0b0-68de44efe94e-config-volume\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.360300 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhvhz\" (UniqueName: \"kubernetes.io/projected/44426e56-e34e-4df8-b0b0-68de44efe94e-kube-api-access-xhvhz\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.361479 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44426e56-e34e-4df8-b0b0-68de44efe94e-config-volume\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.365128 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44426e56-e34e-4df8-b0b0-68de44efe94e-secret-volume\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.378899 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhvhz\" (UniqueName: \"kubernetes.io/projected/44426e56-e34e-4df8-b0b0-68de44efe94e-kube-api-access-xhvhz\") pod \"collect-profiles-29323305-g6rqv\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:00 crc kubenswrapper[5035]: I1002 09:45:00.461570 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:04 crc kubenswrapper[5035]: I1002 09:45:04.785103 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.131:5353: connect: connection refused" Oct 02 09:45:04 crc kubenswrapper[5035]: I1002 09:45:04.785828 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:45:07 crc kubenswrapper[5035]: E1002 09:45:07.859730 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 02 09:45:07 crc kubenswrapper[5035]: E1002 09:45:07.860399 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f9rvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-qf6x5_openstack(830b460b-be63-4b42-a442-80d844ef6908): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:45:07 crc kubenswrapper[5035]: E1002 09:45:07.861666 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-qf6x5" podUID="830b460b-be63-4b42-a442-80d844ef6908" Oct 02 09:45:08 crc kubenswrapper[5035]: E1002 09:45:08.832224 5035 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 02 09:45:08 crc kubenswrapper[5035]: E1002 09:45:08.832714 5035 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mpm2v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7t2tz_openstack(76ce8ea1-cda3-4f68-9570-c6c02b56283f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 09:45:08 crc kubenswrapper[5035]: E1002 09:45:08.834986 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7t2tz" podUID="76ce8ea1-cda3-4f68-9570-c6c02b56283f" Oct 02 09:45:08 crc kubenswrapper[5035]: I1002 09:45:08.861559 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" event={"ID":"9f2fd41d-10e9-4f68-bf8e-8a482b827690","Type":"ContainerDied","Data":"33e38d1624d4780eb8750c0483734da1ccaaaad53bdb9694e78aff9ea4c879b0"} Oct 02 09:45:08 crc kubenswrapper[5035]: I1002 09:45:08.861608 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33e38d1624d4780eb8750c0483734da1ccaaaad53bdb9694e78aff9ea4c879b0" Oct 02 09:45:08 crc kubenswrapper[5035]: I1002 09:45:08.864232 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5cvfn" event={"ID":"942a97bf-0aac-4dec-ba19-34e188adc514","Type":"ContainerDied","Data":"11e295024087e98211503e16609b7e1065a621ee8c83c8f988570a8e8a13a97e"} Oct 02 09:45:08 crc kubenswrapper[5035]: I1002 09:45:08.864274 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11e295024087e98211503e16609b7e1065a621ee8c83c8f988570a8e8a13a97e" Oct 02 09:45:08 crc kubenswrapper[5035]: E1002 09:45:08.868062 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7t2tz" podUID="76ce8ea1-cda3-4f68-9570-c6c02b56283f" Oct 02 09:45:08 crc kubenswrapper[5035]: E1002 09:45:08.895330 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-qf6x5" podUID="830b460b-be63-4b42-a442-80d844ef6908" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.044657 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.057825 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.117346 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-scripts\") pod \"942a97bf-0aac-4dec-ba19-34e188adc514\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.123787 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-sb\") pod \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.123824 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjcss\" (UniqueName: \"kubernetes.io/projected/9f2fd41d-10e9-4f68-bf8e-8a482b827690-kube-api-access-bjcss\") pod \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.123858 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-config\") pod \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.125702 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-fernet-keys\") pod \"942a97bf-0aac-4dec-ba19-34e188adc514\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.125727 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-swift-storage-0\") pod \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.125757 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-svc\") pod \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.125800 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-nb\") pod \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\" (UID: \"9f2fd41d-10e9-4f68-bf8e-8a482b827690\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.125842 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-credential-keys\") pod \"942a97bf-0aac-4dec-ba19-34e188adc514\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.125893 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-combined-ca-bundle\") pod \"942a97bf-0aac-4dec-ba19-34e188adc514\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.125988 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-config-data\") pod \"942a97bf-0aac-4dec-ba19-34e188adc514\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.126024 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jczv9\" (UniqueName: \"kubernetes.io/projected/942a97bf-0aac-4dec-ba19-34e188adc514-kube-api-access-jczv9\") pod \"942a97bf-0aac-4dec-ba19-34e188adc514\" (UID: \"942a97bf-0aac-4dec-ba19-34e188adc514\") " Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.129013 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-scripts" (OuterVolumeSpecName: "scripts") pod "942a97bf-0aac-4dec-ba19-34e188adc514" (UID: "942a97bf-0aac-4dec-ba19-34e188adc514"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.150041 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f2fd41d-10e9-4f68-bf8e-8a482b827690-kube-api-access-bjcss" (OuterVolumeSpecName: "kube-api-access-bjcss") pod "9f2fd41d-10e9-4f68-bf8e-8a482b827690" (UID: "9f2fd41d-10e9-4f68-bf8e-8a482b827690"). InnerVolumeSpecName "kube-api-access-bjcss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.158928 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "942a97bf-0aac-4dec-ba19-34e188adc514" (UID: "942a97bf-0aac-4dec-ba19-34e188adc514"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.161169 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/942a97bf-0aac-4dec-ba19-34e188adc514-kube-api-access-jczv9" (OuterVolumeSpecName: "kube-api-access-jczv9") pod "942a97bf-0aac-4dec-ba19-34e188adc514" (UID: "942a97bf-0aac-4dec-ba19-34e188adc514"). InnerVolumeSpecName "kube-api-access-jczv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.165645 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "942a97bf-0aac-4dec-ba19-34e188adc514" (UID: "942a97bf-0aac-4dec-ba19-34e188adc514"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.205965 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-config-data" (OuterVolumeSpecName: "config-data") pod "942a97bf-0aac-4dec-ba19-34e188adc514" (UID: "942a97bf-0aac-4dec-ba19-34e188adc514"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.227649 5035 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.227681 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.227691 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jczv9\" (UniqueName: \"kubernetes.io/projected/942a97bf-0aac-4dec-ba19-34e188adc514-kube-api-access-jczv9\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.227700 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.227709 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjcss\" (UniqueName: \"kubernetes.io/projected/9f2fd41d-10e9-4f68-bf8e-8a482b827690-kube-api-access-bjcss\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.227718 5035 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.234356 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-config" (OuterVolumeSpecName: "config") pod "9f2fd41d-10e9-4f68-bf8e-8a482b827690" (UID: "9f2fd41d-10e9-4f68-bf8e-8a482b827690"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.234682 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "942a97bf-0aac-4dec-ba19-34e188adc514" (UID: "942a97bf-0aac-4dec-ba19-34e188adc514"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.239279 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9f2fd41d-10e9-4f68-bf8e-8a482b827690" (UID: "9f2fd41d-10e9-4f68-bf8e-8a482b827690"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.244605 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9f2fd41d-10e9-4f68-bf8e-8a482b827690" (UID: "9f2fd41d-10e9-4f68-bf8e-8a482b827690"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.250372 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9f2fd41d-10e9-4f68-bf8e-8a482b827690" (UID: "9f2fd41d-10e9-4f68-bf8e-8a482b827690"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.251155 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9f2fd41d-10e9-4f68-bf8e-8a482b827690" (UID: "9f2fd41d-10e9-4f68-bf8e-8a482b827690"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.263880 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-8c07-account-create-9t9l8"] Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.331632 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942a97bf-0aac-4dec-ba19-34e188adc514-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.331685 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.331706 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.331739 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.331758 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.331787 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9f2fd41d-10e9-4f68-bf8e-8a482b827690-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.441491 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-286pl"] Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.514962 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:45:09 crc kubenswrapper[5035]: W1002 09:45:09.516104 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e598700_b625_46eb_9807_9f81dc5e7bc5.slice/crio-37f8c4d32a752dba3cc54251a60218fbec1f9078b7af91ba51d664c61252b02d WatchSource:0}: Error finding container 37f8c4d32a752dba3cc54251a60218fbec1f9078b7af91ba51d664c61252b02d: Status 404 returned error can't find the container with id 37f8c4d32a752dba3cc54251a60218fbec1f9078b7af91ba51d664c61252b02d Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.676313 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv"] Oct 02 09:45:09 crc kubenswrapper[5035]: W1002 09:45:09.697244 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44426e56_e34e_4df8_b0b0_68de44efe94e.slice/crio-c7fd9332e3ce719b9abd2046aae472a8284d7386df1b2b83c7e174b8a7670888 WatchSource:0}: Error finding container c7fd9332e3ce719b9abd2046aae472a8284d7386df1b2b83c7e174b8a7670888: Status 404 returned error can't find the container with id c7fd9332e3ce719b9abd2046aae472a8284d7386df1b2b83c7e174b8a7670888 Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.886916 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-knpjm" event={"ID":"136b4da6-5c57-4a29-af5a-afe32c5f552f","Type":"ContainerStarted","Data":"02b19eba82aa11ef3716fd2aa1d250359f0f75877f48cae54ce55304d1c6e091"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.889858 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" event={"ID":"44426e56-e34e-4df8-b0b0-68de44efe94e","Type":"ContainerStarted","Data":"c7fd9332e3ce719b9abd2046aae472a8284d7386df1b2b83c7e174b8a7670888"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.892868 5035 generic.go:334] "Generic (PLEG): container finished" podID="2887e3dc-2301-42c6-ac2d-725adefd5fcf" containerID="af2fc1b21d33c06faf5156a6bb204aed8e2faadf70b5f77609fdd6dd962f9909" exitCode=0 Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.892988 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-8c07-account-create-9t9l8" event={"ID":"2887e3dc-2301-42c6-ac2d-725adefd5fcf","Type":"ContainerDied","Data":"af2fc1b21d33c06faf5156a6bb204aed8e2faadf70b5f77609fdd6dd962f9909"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.893020 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-8c07-account-create-9t9l8" event={"ID":"2887e3dc-2301-42c6-ac2d-725adefd5fcf","Type":"ContainerStarted","Data":"0e33b719aa0015a1cb6234a1977ae8ce399bb040310876032d9a7d20863d8402"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.899563 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e598700-b625-46eb-9807-9f81dc5e7bc5","Type":"ContainerStarted","Data":"37f8c4d32a752dba3cc54251a60218fbec1f9078b7af91ba51d664c61252b02d"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.901710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerStarted","Data":"17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.910669 5035 generic.go:334] "Generic (PLEG): container finished" podID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerID="0d0a7a63acd78ac35565d708cfecfec5228b463d489da782ac260ec7c5e708a8" exitCode=0 Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.910856 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-2rxp7" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.911559 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" event={"ID":"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45","Type":"ContainerDied","Data":"0d0a7a63acd78ac35565d708cfecfec5228b463d489da782ac260ec7c5e708a8"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.911592 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" event={"ID":"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45","Type":"ContainerStarted","Data":"71d46c208f70aaeb8f8a9d17c92821c2c3991c4defacdc07151ab15238017960"} Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.911973 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5cvfn" Oct 02 09:45:09 crc kubenswrapper[5035]: I1002 09:45:09.913371 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-knpjm" podStartSLOduration=3.185377173 podStartE2EDuration="30.9133431s" podCreationTimestamp="2025-10-02 09:44:39 +0000 UTC" firstStartedPulling="2025-10-02 09:44:41.213643402 +0000 UTC m=+1046.569987427" lastFinishedPulling="2025-10-02 09:45:08.941609309 +0000 UTC m=+1074.297953354" observedRunningTime="2025-10-02 09:45:09.90645213 +0000 UTC m=+1075.262796155" watchObservedRunningTime="2025-10-02 09:45:09.9133431 +0000 UTC m=+1075.269687125" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.134926 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5cvfn"] Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.149147 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5cvfn"] Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.184216 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="942a97bf-0aac-4dec-ba19-34e188adc514" path="/var/lib/kubelet/pods/942a97bf-0aac-4dec-ba19-34e188adc514/volumes" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.185045 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-2rxp7"] Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.190921 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-2rxp7"] Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.247219 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xb6lf"] Oct 02 09:45:10 crc kubenswrapper[5035]: E1002 09:45:10.247774 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="init" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.247803 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="init" Oct 02 09:45:10 crc kubenswrapper[5035]: E1002 09:45:10.247823 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="dnsmasq-dns" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.247832 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="dnsmasq-dns" Oct 02 09:45:10 crc kubenswrapper[5035]: E1002 09:45:10.247851 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="942a97bf-0aac-4dec-ba19-34e188adc514" containerName="keystone-bootstrap" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.247859 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="942a97bf-0aac-4dec-ba19-34e188adc514" containerName="keystone-bootstrap" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.248214 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="942a97bf-0aac-4dec-ba19-34e188adc514" containerName="keystone-bootstrap" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.248240 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" containerName="dnsmasq-dns" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.248825 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.254245 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.254447 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.254586 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fj9kx" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.254685 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.256950 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xb6lf"] Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.387495 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-fernet-keys\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.387571 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-config-data\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.387650 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-scripts\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.387725 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-credential-keys\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.387852 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2vgk\" (UniqueName: \"kubernetes.io/projected/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-kube-api-access-r2vgk\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.388074 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-combined-ca-bundle\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.490100 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-combined-ca-bundle\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.490168 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-fernet-keys\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.490194 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-config-data\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.490231 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-scripts\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.490284 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-credential-keys\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.490305 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2vgk\" (UniqueName: \"kubernetes.io/projected/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-kube-api-access-r2vgk\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.498480 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-combined-ca-bundle\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.498723 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-fernet-keys\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.499161 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-credential-keys\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.500859 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-scripts\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.504628 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-config-data\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.510650 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2vgk\" (UniqueName: \"kubernetes.io/projected/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-kube-api-access-r2vgk\") pod \"keystone-bootstrap-xb6lf\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.575610 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.649877 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.934730 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce7048e6-3e78-4655-b89a-9f010dcec708","Type":"ContainerStarted","Data":"a51819184ed5687fabb590b4d418a1dd5038e7db762c760311f10d60dce35c37"} Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.939148 5035 generic.go:334] "Generic (PLEG): container finished" podID="44426e56-e34e-4df8-b0b0-68de44efe94e" containerID="8e9fb8f17e4f1615079782805c6bbb957e549321278d0659b0f1ee2b74acddff" exitCode=0 Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.939214 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" event={"ID":"44426e56-e34e-4df8-b0b0-68de44efe94e","Type":"ContainerDied","Data":"8e9fb8f17e4f1615079782805c6bbb957e549321278d0659b0f1ee2b74acddff"} Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.940749 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e598700-b625-46eb-9807-9f81dc5e7bc5","Type":"ContainerStarted","Data":"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383"} Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.943932 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerStarted","Data":"04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673"} Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.953870 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" event={"ID":"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45","Type":"ContainerStarted","Data":"bb0957b6c8828276a9c0dbdb0b34512261a2a5c3c1c350435b3a4024026c18f2"} Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.955104 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:45:10 crc kubenswrapper[5035]: I1002 09:45:10.979304 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" podStartSLOduration=19.979288646 podStartE2EDuration="19.979288646s" podCreationTimestamp="2025-10-02 09:44:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:10.977297988 +0000 UTC m=+1076.333642023" watchObservedRunningTime="2025-10-02 09:45:10.979288646 +0000 UTC m=+1076.335632671" Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.079324 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xb6lf"] Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.365612 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8c07-account-create-9t9l8" Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.523583 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5ksm\" (UniqueName: \"kubernetes.io/projected/2887e3dc-2301-42c6-ac2d-725adefd5fcf-kube-api-access-w5ksm\") pod \"2887e3dc-2301-42c6-ac2d-725adefd5fcf\" (UID: \"2887e3dc-2301-42c6-ac2d-725adefd5fcf\") " Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.527270 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2887e3dc-2301-42c6-ac2d-725adefd5fcf-kube-api-access-w5ksm" (OuterVolumeSpecName: "kube-api-access-w5ksm") pod "2887e3dc-2301-42c6-ac2d-725adefd5fcf" (UID: "2887e3dc-2301-42c6-ac2d-725adefd5fcf"). InnerVolumeSpecName "kube-api-access-w5ksm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.626386 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5ksm\" (UniqueName: \"kubernetes.io/projected/2887e3dc-2301-42c6-ac2d-725adefd5fcf-kube-api-access-w5ksm\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.965748 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xb6lf" event={"ID":"2ad8a14e-58b5-4192-acf8-972ffdbd8e27","Type":"ContainerStarted","Data":"16ad5aded9c7a6befe8479af76936081fa37f639cc6ade8422791922b3bf1318"} Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.967289 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xb6lf" event={"ID":"2ad8a14e-58b5-4192-acf8-972ffdbd8e27","Type":"ContainerStarted","Data":"17d5440a87165c8c0a691ee96c6243d513d632697ffd4c861e0d0f1490b718c2"} Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.970502 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-8c07-account-create-9t9l8" event={"ID":"2887e3dc-2301-42c6-ac2d-725adefd5fcf","Type":"ContainerDied","Data":"0e33b719aa0015a1cb6234a1977ae8ce399bb040310876032d9a7d20863d8402"} Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.970570 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e33b719aa0015a1cb6234a1977ae8ce399bb040310876032d9a7d20863d8402" Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.970658 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8c07-account-create-9t9l8" Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.990085 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e598700-b625-46eb-9807-9f81dc5e7bc5","Type":"ContainerStarted","Data":"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22"} Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.990227 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-log" containerID="cri-o://04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383" gracePeriod=30 Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.990303 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-httpd" containerID="cri-o://80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22" gracePeriod=30 Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.995165 5035 generic.go:334] "Generic (PLEG): container finished" podID="136b4da6-5c57-4a29-af5a-afe32c5f552f" containerID="02b19eba82aa11ef3716fd2aa1d250359f0f75877f48cae54ce55304d1c6e091" exitCode=0 Oct 02 09:45:11 crc kubenswrapper[5035]: I1002 09:45:11.995236 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-knpjm" event={"ID":"136b4da6-5c57-4a29-af5a-afe32c5f552f","Type":"ContainerDied","Data":"02b19eba82aa11ef3716fd2aa1d250359f0f75877f48cae54ce55304d1c6e091"} Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.006442 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce7048e6-3e78-4655-b89a-9f010dcec708","Type":"ContainerStarted","Data":"f6d79d95ef220440c49fd6178448d63000f20f5d772de27c03380b1dfe210aa3"} Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.017571 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xb6lf" podStartSLOduration=2.017549905 podStartE2EDuration="2.017549905s" podCreationTimestamp="2025-10-02 09:45:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:11.980678441 +0000 UTC m=+1077.337022466" watchObservedRunningTime="2025-10-02 09:45:12.017549905 +0000 UTC m=+1077.373893940" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.020775 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.020761359 podStartE2EDuration="21.020761359s" podCreationTimestamp="2025-10-02 09:44:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:12.015772863 +0000 UTC m=+1077.372116878" watchObservedRunningTime="2025-10-02 09:45:12.020761359 +0000 UTC m=+1077.377105384" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.178825 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f2fd41d-10e9-4f68-bf8e-8a482b827690" path="/var/lib/kubelet/pods/9f2fd41d-10e9-4f68-bf8e-8a482b827690/volumes" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.378618 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.543115 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44426e56-e34e-4df8-b0b0-68de44efe94e-config-volume\") pod \"44426e56-e34e-4df8-b0b0-68de44efe94e\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.543260 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44426e56-e34e-4df8-b0b0-68de44efe94e-secret-volume\") pod \"44426e56-e34e-4df8-b0b0-68de44efe94e\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.543416 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhvhz\" (UniqueName: \"kubernetes.io/projected/44426e56-e34e-4df8-b0b0-68de44efe94e-kube-api-access-xhvhz\") pod \"44426e56-e34e-4df8-b0b0-68de44efe94e\" (UID: \"44426e56-e34e-4df8-b0b0-68de44efe94e\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.543745 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44426e56-e34e-4df8-b0b0-68de44efe94e-config-volume" (OuterVolumeSpecName: "config-volume") pod "44426e56-e34e-4df8-b0b0-68de44efe94e" (UID: "44426e56-e34e-4df8-b0b0-68de44efe94e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.543894 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/44426e56-e34e-4df8-b0b0-68de44efe94e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.551009 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44426e56-e34e-4df8-b0b0-68de44efe94e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "44426e56-e34e-4df8-b0b0-68de44efe94e" (UID: "44426e56-e34e-4df8-b0b0-68de44efe94e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.552148 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44426e56-e34e-4df8-b0b0-68de44efe94e-kube-api-access-xhvhz" (OuterVolumeSpecName: "kube-api-access-xhvhz") pod "44426e56-e34e-4df8-b0b0-68de44efe94e" (UID: "44426e56-e34e-4df8-b0b0-68de44efe94e"). InnerVolumeSpecName "kube-api-access-xhvhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.630139 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.645805 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/44426e56-e34e-4df8-b0b0-68de44efe94e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.645855 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhvhz\" (UniqueName: \"kubernetes.io/projected/44426e56-e34e-4df8-b0b0-68de44efe94e-kube-api-access-xhvhz\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.746758 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"6e598700-b625-46eb-9807-9f81dc5e7bc5\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747090 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-logs\") pod \"6e598700-b625-46eb-9807-9f81dc5e7bc5\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747128 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-httpd-run\") pod \"6e598700-b625-46eb-9807-9f81dc5e7bc5\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747230 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-scripts\") pod \"6e598700-b625-46eb-9807-9f81dc5e7bc5\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747277 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-config-data\") pod \"6e598700-b625-46eb-9807-9f81dc5e7bc5\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747608 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6e598700-b625-46eb-9807-9f81dc5e7bc5" (UID: "6e598700-b625-46eb-9807-9f81dc5e7bc5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747719 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzw7f\" (UniqueName: \"kubernetes.io/projected/6e598700-b625-46eb-9807-9f81dc5e7bc5-kube-api-access-nzw7f\") pod \"6e598700-b625-46eb-9807-9f81dc5e7bc5\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747777 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-combined-ca-bundle\") pod \"6e598700-b625-46eb-9807-9f81dc5e7bc5\" (UID: \"6e598700-b625-46eb-9807-9f81dc5e7bc5\") " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.747719 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-logs" (OuterVolumeSpecName: "logs") pod "6e598700-b625-46eb-9807-9f81dc5e7bc5" (UID: "6e598700-b625-46eb-9807-9f81dc5e7bc5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.748588 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.748613 5035 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e598700-b625-46eb-9807-9f81dc5e7bc5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.753646 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e598700-b625-46eb-9807-9f81dc5e7bc5-kube-api-access-nzw7f" (OuterVolumeSpecName: "kube-api-access-nzw7f") pod "6e598700-b625-46eb-9807-9f81dc5e7bc5" (UID: "6e598700-b625-46eb-9807-9f81dc5e7bc5"). InnerVolumeSpecName "kube-api-access-nzw7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.754131 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "6e598700-b625-46eb-9807-9f81dc5e7bc5" (UID: "6e598700-b625-46eb-9807-9f81dc5e7bc5"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.755786 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-scripts" (OuterVolumeSpecName: "scripts") pod "6e598700-b625-46eb-9807-9f81dc5e7bc5" (UID: "6e598700-b625-46eb-9807-9f81dc5e7bc5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.788039 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e598700-b625-46eb-9807-9f81dc5e7bc5" (UID: "6e598700-b625-46eb-9807-9f81dc5e7bc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.807400 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-config-data" (OuterVolumeSpecName: "config-data") pod "6e598700-b625-46eb-9807-9f81dc5e7bc5" (UID: "6e598700-b625-46eb-9807-9f81dc5e7bc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.859203 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzw7f\" (UniqueName: \"kubernetes.io/projected/6e598700-b625-46eb-9807-9f81dc5e7bc5-kube-api-access-nzw7f\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.859243 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.859275 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.859284 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.859293 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e598700-b625-46eb-9807-9f81dc5e7bc5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.877870 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 02 09:45:12 crc kubenswrapper[5035]: I1002 09:45:12.960927 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.022395 5035 generic.go:334] "Generic (PLEG): container finished" podID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerID="80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22" exitCode=0 Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.022428 5035 generic.go:334] "Generic (PLEG): container finished" podID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerID="04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383" exitCode=143 Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.022454 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e598700-b625-46eb-9807-9f81dc5e7bc5","Type":"ContainerDied","Data":"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22"} Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.022556 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.022613 5035 scope.go:117] "RemoveContainer" containerID="80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.022637 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e598700-b625-46eb-9807-9f81dc5e7bc5","Type":"ContainerDied","Data":"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383"} Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.022663 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6e598700-b625-46eb-9807-9f81dc5e7bc5","Type":"ContainerDied","Data":"37f8c4d32a752dba3cc54251a60218fbec1f9078b7af91ba51d664c61252b02d"} Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.032731 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.033080 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323305-g6rqv" event={"ID":"44426e56-e34e-4df8-b0b0-68de44efe94e","Type":"ContainerDied","Data":"c7fd9332e3ce719b9abd2046aae472a8284d7386df1b2b83c7e174b8a7670888"} Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.033144 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7fd9332e3ce719b9abd2046aae472a8284d7386df1b2b83c7e174b8a7670888" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.041216 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-log" containerID="cri-o://f6d79d95ef220440c49fd6178448d63000f20f5d772de27c03380b1dfe210aa3" gracePeriod=30 Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.041435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce7048e6-3e78-4655-b89a-9f010dcec708","Type":"ContainerStarted","Data":"3ffd2a9e567ac65f44500bafeda0b9f77585a24cf6c1618dc1f27db173473226"} Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.041697 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-httpd" containerID="cri-o://3ffd2a9e567ac65f44500bafeda0b9f77585a24cf6c1618dc1f27db173473226" gracePeriod=30 Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.074838 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=23.074819787 podStartE2EDuration="23.074819787s" podCreationTimestamp="2025-10-02 09:44:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:13.064592509 +0000 UTC m=+1078.420936534" watchObservedRunningTime="2025-10-02 09:45:13.074819787 +0000 UTC m=+1078.431163812" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.094858 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.117389 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.131927 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:45:13 crc kubenswrapper[5035]: E1002 09:45:13.132376 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44426e56-e34e-4df8-b0b0-68de44efe94e" containerName="collect-profiles" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132395 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="44426e56-e34e-4df8-b0b0-68de44efe94e" containerName="collect-profiles" Oct 02 09:45:13 crc kubenswrapper[5035]: E1002 09:45:13.132417 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2887e3dc-2301-42c6-ac2d-725adefd5fcf" containerName="mariadb-account-create" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132423 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2887e3dc-2301-42c6-ac2d-725adefd5fcf" containerName="mariadb-account-create" Oct 02 09:45:13 crc kubenswrapper[5035]: E1002 09:45:13.132439 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-log" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132444 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-log" Oct 02 09:45:13 crc kubenswrapper[5035]: E1002 09:45:13.132454 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-httpd" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132461 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-httpd" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132642 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="44426e56-e34e-4df8-b0b0-68de44efe94e" containerName="collect-profiles" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132659 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2887e3dc-2301-42c6-ac2d-725adefd5fcf" containerName="mariadb-account-create" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132671 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-log" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.132687 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" containerName="glance-httpd" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.133565 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.142741 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.175025 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.176263 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.276438 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.276478 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.276524 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.276555 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69wj2\" (UniqueName: \"kubernetes.io/projected/f42b1109-c1b4-4a08-a8d3-03cc50774a90-kube-api-access-69wj2\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.276585 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-logs\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.277021 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.277097 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.277165 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.378781 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379108 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379131 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379160 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379176 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69wj2\" (UniqueName: \"kubernetes.io/projected/f42b1109-c1b4-4a08-a8d3-03cc50774a90-kube-api-access-69wj2\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379195 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-logs\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379247 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379264 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379588 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.379704 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.380132 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-logs\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.384831 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.384904 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.385044 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.397205 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.398119 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69wj2\" (UniqueName: \"kubernetes.io/projected/f42b1109-c1b4-4a08-a8d3-03cc50774a90-kube-api-access-69wj2\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.409033 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:45:13 crc kubenswrapper[5035]: I1002 09:45:13.502628 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:14 crc kubenswrapper[5035]: I1002 09:45:14.050719 5035 generic.go:334] "Generic (PLEG): container finished" podID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerID="3ffd2a9e567ac65f44500bafeda0b9f77585a24cf6c1618dc1f27db173473226" exitCode=0 Oct 02 09:45:14 crc kubenswrapper[5035]: I1002 09:45:14.050748 5035 generic.go:334] "Generic (PLEG): container finished" podID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerID="f6d79d95ef220440c49fd6178448d63000f20f5d772de27c03380b1dfe210aa3" exitCode=143 Oct 02 09:45:14 crc kubenswrapper[5035]: I1002 09:45:14.050820 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce7048e6-3e78-4655-b89a-9f010dcec708","Type":"ContainerDied","Data":"3ffd2a9e567ac65f44500bafeda0b9f77585a24cf6c1618dc1f27db173473226"} Oct 02 09:45:14 crc kubenswrapper[5035]: I1002 09:45:14.050860 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce7048e6-3e78-4655-b89a-9f010dcec708","Type":"ContainerDied","Data":"f6d79d95ef220440c49fd6178448d63000f20f5d772de27c03380b1dfe210aa3"} Oct 02 09:45:14 crc kubenswrapper[5035]: I1002 09:45:14.173656 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e598700-b625-46eb-9807-9f81dc5e7bc5" path="/var/lib/kubelet/pods/6e598700-b625-46eb-9807-9f81dc5e7bc5/volumes" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.070231 5035 generic.go:334] "Generic (PLEG): container finished" podID="2ad8a14e-58b5-4192-acf8-972ffdbd8e27" containerID="16ad5aded9c7a6befe8479af76936081fa37f639cc6ade8422791922b3bf1318" exitCode=0 Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.070270 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xb6lf" event={"ID":"2ad8a14e-58b5-4192-acf8-972ffdbd8e27","Type":"ContainerDied","Data":"16ad5aded9c7a6befe8479af76936081fa37f639cc6ade8422791922b3bf1318"} Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.125509 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-sync-klp2h"] Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.127023 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.129602 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-8t5qg" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.129636 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-scripts" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.129683 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.142766 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-klp2h"] Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.210496 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/afad3fdc-5375-4bba-9832-f94381ba82aa-etc-podinfo\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.210836 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spqqn\" (UniqueName: \"kubernetes.io/projected/afad3fdc-5375-4bba-9832-f94381ba82aa-kube-api-access-spqqn\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.210931 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.211011 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-combined-ca-bundle\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.211202 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data-merged\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.211413 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-scripts\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.313752 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/afad3fdc-5375-4bba-9832-f94381ba82aa-etc-podinfo\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.313833 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spqqn\" (UniqueName: \"kubernetes.io/projected/afad3fdc-5375-4bba-9832-f94381ba82aa-kube-api-access-spqqn\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.313868 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.313888 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-combined-ca-bundle\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.313928 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data-merged\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.314030 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-scripts\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.316655 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data-merged\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.323834 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/afad3fdc-5375-4bba-9832-f94381ba82aa-etc-podinfo\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.324444 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-scripts\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.326056 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-combined-ca-bundle\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.333651 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.333845 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spqqn\" (UniqueName: \"kubernetes.io/projected/afad3fdc-5375-4bba-9832-f94381ba82aa-kube-api-access-spqqn\") pod \"ironic-db-sync-klp2h\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:15 crc kubenswrapper[5035]: I1002 09:45:15.455139 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:16 crc kubenswrapper[5035]: I1002 09:45:16.430797 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:45:16 crc kubenswrapper[5035]: I1002 09:45:16.530365 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-h8tfl"] Oct 02 09:45:16 crc kubenswrapper[5035]: I1002 09:45:16.534597 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" podUID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerName="dnsmasq-dns" containerID="cri-o://116c59c19ce9de2fa774d4767abea7fe77a099dde4c885273caa655d9818a6fc" gracePeriod=10 Oct 02 09:45:17 crc kubenswrapper[5035]: I1002 09:45:17.091854 5035 generic.go:334] "Generic (PLEG): container finished" podID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerID="116c59c19ce9de2fa774d4767abea7fe77a099dde4c885273caa655d9818a6fc" exitCode=0 Oct 02 09:45:17 crc kubenswrapper[5035]: I1002 09:45:17.091907 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" event={"ID":"fa2948a0-da02-4c95-bc2b-d9789564c9e7","Type":"ContainerDied","Data":"116c59c19ce9de2fa774d4767abea7fe77a099dde4c885273caa655d9818a6fc"} Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.026576 5035 scope.go:117] "RemoveContainer" containerID="04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.203778 5035 scope.go:117] "RemoveContainer" containerID="80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22" Oct 02 09:45:18 crc kubenswrapper[5035]: E1002 09:45:18.204198 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22\": container with ID starting with 80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22 not found: ID does not exist" containerID="80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.204243 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22"} err="failed to get container status \"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22\": rpc error: code = NotFound desc = could not find container \"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22\": container with ID starting with 80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22 not found: ID does not exist" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.204278 5035 scope.go:117] "RemoveContainer" containerID="04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383" Oct 02 09:45:18 crc kubenswrapper[5035]: E1002 09:45:18.204697 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383\": container with ID starting with 04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383 not found: ID does not exist" containerID="04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.204731 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383"} err="failed to get container status \"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383\": rpc error: code = NotFound desc = could not find container \"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383\": container with ID starting with 04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383 not found: ID does not exist" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.204751 5035 scope.go:117] "RemoveContainer" containerID="80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.205109 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22"} err="failed to get container status \"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22\": rpc error: code = NotFound desc = could not find container \"80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22\": container with ID starting with 80e286e5d7be723e484477a1d31172e59b3290323fbbe9acb67ecf32d4973e22 not found: ID does not exist" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.205142 5035 scope.go:117] "RemoveContainer" containerID="04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.205472 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383"} err="failed to get container status \"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383\": rpc error: code = NotFound desc = could not find container \"04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383\": container with ID starting with 04e0d939a11207ace85547d6355e139312b63b7315afbaf7b8dd2efc4afea383 not found: ID does not exist" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.315277 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.332411 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-knpjm" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.479864 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-combined-ca-bundle\") pod \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.479900 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-scripts\") pod \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.479970 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-config-data\") pod \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.479990 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-config-data\") pod \"136b4da6-5c57-4a29-af5a-afe32c5f552f\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480113 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-scripts\") pod \"136b4da6-5c57-4a29-af5a-afe32c5f552f\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480140 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136b4da6-5c57-4a29-af5a-afe32c5f552f-logs\") pod \"136b4da6-5c57-4a29-af5a-afe32c5f552f\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480160 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2k7k\" (UniqueName: \"kubernetes.io/projected/136b4da6-5c57-4a29-af5a-afe32c5f552f-kube-api-access-t2k7k\") pod \"136b4da6-5c57-4a29-af5a-afe32c5f552f\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480194 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-fernet-keys\") pod \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480212 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-combined-ca-bundle\") pod \"136b4da6-5c57-4a29-af5a-afe32c5f552f\" (UID: \"136b4da6-5c57-4a29-af5a-afe32c5f552f\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480252 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2vgk\" (UniqueName: \"kubernetes.io/projected/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-kube-api-access-r2vgk\") pod \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480291 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-credential-keys\") pod \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\" (UID: \"2ad8a14e-58b5-4192-acf8-972ffdbd8e27\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.480990 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/136b4da6-5c57-4a29-af5a-afe32c5f552f-logs" (OuterVolumeSpecName: "logs") pod "136b4da6-5c57-4a29-af5a-afe32c5f552f" (UID: "136b4da6-5c57-4a29-af5a-afe32c5f552f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.485971 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-kube-api-access-r2vgk" (OuterVolumeSpecName: "kube-api-access-r2vgk") pod "2ad8a14e-58b5-4192-acf8-972ffdbd8e27" (UID: "2ad8a14e-58b5-4192-acf8-972ffdbd8e27"). InnerVolumeSpecName "kube-api-access-r2vgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.489074 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2ad8a14e-58b5-4192-acf8-972ffdbd8e27" (UID: "2ad8a14e-58b5-4192-acf8-972ffdbd8e27"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.490965 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-scripts" (OuterVolumeSpecName: "scripts") pod "136b4da6-5c57-4a29-af5a-afe32c5f552f" (UID: "136b4da6-5c57-4a29-af5a-afe32c5f552f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.491067 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/136b4da6-5c57-4a29-af5a-afe32c5f552f-kube-api-access-t2k7k" (OuterVolumeSpecName: "kube-api-access-t2k7k") pod "136b4da6-5c57-4a29-af5a-afe32c5f552f" (UID: "136b4da6-5c57-4a29-af5a-afe32c5f552f"). InnerVolumeSpecName "kube-api-access-t2k7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.506377 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2ad8a14e-58b5-4192-acf8-972ffdbd8e27" (UID: "2ad8a14e-58b5-4192-acf8-972ffdbd8e27"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.506579 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-scripts" (OuterVolumeSpecName: "scripts") pod "2ad8a14e-58b5-4192-acf8-972ffdbd8e27" (UID: "2ad8a14e-58b5-4192-acf8-972ffdbd8e27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.544247 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "136b4da6-5c57-4a29-af5a-afe32c5f552f" (UID: "136b4da6-5c57-4a29-af5a-afe32c5f552f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.564232 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-config-data" (OuterVolumeSpecName: "config-data") pod "2ad8a14e-58b5-4192-acf8-972ffdbd8e27" (UID: "2ad8a14e-58b5-4192-acf8-972ffdbd8e27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.571787 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2ad8a14e-58b5-4192-acf8-972ffdbd8e27" (UID: "2ad8a14e-58b5-4192-acf8-972ffdbd8e27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585252 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585295 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585306 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/136b4da6-5c57-4a29-af5a-afe32c5f552f-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585317 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2k7k\" (UniqueName: \"kubernetes.io/projected/136b4da6-5c57-4a29-af5a-afe32c5f552f-kube-api-access-t2k7k\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585331 5035 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585341 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585353 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2vgk\" (UniqueName: \"kubernetes.io/projected/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-kube-api-access-r2vgk\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585366 5035 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585378 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.585388 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2ad8a14e-58b5-4192-acf8-972ffdbd8e27-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.588087 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-config-data" (OuterVolumeSpecName: "config-data") pod "136b4da6-5c57-4a29-af5a-afe32c5f552f" (UID: "136b4da6-5c57-4a29-af5a-afe32c5f552f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.611807 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.649980 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.688080 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/136b4da6-5c57-4a29-af5a-afe32c5f552f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.778869 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-klp2h"] Oct 02 09:45:18 crc kubenswrapper[5035]: W1002 09:45:18.781840 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafad3fdc_5375_4bba_9832_f94381ba82aa.slice/crio-c27323c7e57c6556dc856f0f30e811673f692f7309ba9c7db35a02c8e13399ef WatchSource:0}: Error finding container c27323c7e57c6556dc856f0f30e811673f692f7309ba9c7db35a02c8e13399ef: Status 404 returned error can't find the container with id c27323c7e57c6556dc856f0f30e811673f692f7309ba9c7db35a02c8e13399ef Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789009 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-combined-ca-bundle\") pod \"ce7048e6-3e78-4655-b89a-9f010dcec708\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789103 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-swift-storage-0\") pod \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789173 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-scripts\") pod \"ce7048e6-3e78-4655-b89a-9f010dcec708\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789198 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-sb\") pod \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789222 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dt5vs\" (UniqueName: \"kubernetes.io/projected/ce7048e6-3e78-4655-b89a-9f010dcec708-kube-api-access-dt5vs\") pod \"ce7048e6-3e78-4655-b89a-9f010dcec708\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789300 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-config-data\") pod \"ce7048e6-3e78-4655-b89a-9f010dcec708\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789331 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95x2t\" (UniqueName: \"kubernetes.io/projected/fa2948a0-da02-4c95-bc2b-d9789564c9e7-kube-api-access-95x2t\") pod \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789356 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-logs\") pod \"ce7048e6-3e78-4655-b89a-9f010dcec708\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789384 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-svc\") pod \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789422 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-nb\") pod \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789470 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ce7048e6-3e78-4655-b89a-9f010dcec708\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789504 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-config\") pod \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\" (UID: \"fa2948a0-da02-4c95-bc2b-d9789564c9e7\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.789574 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-httpd-run\") pod \"ce7048e6-3e78-4655-b89a-9f010dcec708\" (UID: \"ce7048e6-3e78-4655-b89a-9f010dcec708\") " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.790231 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ce7048e6-3e78-4655-b89a-9f010dcec708" (UID: "ce7048e6-3e78-4655-b89a-9f010dcec708"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.790914 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-logs" (OuterVolumeSpecName: "logs") pod "ce7048e6-3e78-4655-b89a-9f010dcec708" (UID: "ce7048e6-3e78-4655-b89a-9f010dcec708"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.794140 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7048e6-3e78-4655-b89a-9f010dcec708-kube-api-access-dt5vs" (OuterVolumeSpecName: "kube-api-access-dt5vs") pod "ce7048e6-3e78-4655-b89a-9f010dcec708" (UID: "ce7048e6-3e78-4655-b89a-9f010dcec708"). InnerVolumeSpecName "kube-api-access-dt5vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.794259 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa2948a0-da02-4c95-bc2b-d9789564c9e7-kube-api-access-95x2t" (OuterVolumeSpecName: "kube-api-access-95x2t") pod "fa2948a0-da02-4c95-bc2b-d9789564c9e7" (UID: "fa2948a0-da02-4c95-bc2b-d9789564c9e7"). InnerVolumeSpecName "kube-api-access-95x2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.794873 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-scripts" (OuterVolumeSpecName: "scripts") pod "ce7048e6-3e78-4655-b89a-9f010dcec708" (UID: "ce7048e6-3e78-4655-b89a-9f010dcec708"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.796863 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "ce7048e6-3e78-4655-b89a-9f010dcec708" (UID: "ce7048e6-3e78-4655-b89a-9f010dcec708"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.849305 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce7048e6-3e78-4655-b89a-9f010dcec708" (UID: "ce7048e6-3e78-4655-b89a-9f010dcec708"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.851923 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fa2948a0-da02-4c95-bc2b-d9789564c9e7" (UID: "fa2948a0-da02-4c95-bc2b-d9789564c9e7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.853053 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fa2948a0-da02-4c95-bc2b-d9789564c9e7" (UID: "fa2948a0-da02-4c95-bc2b-d9789564c9e7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.863207 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-config-data" (OuterVolumeSpecName: "config-data") pod "ce7048e6-3e78-4655-b89a-9f010dcec708" (UID: "ce7048e6-3e78-4655-b89a-9f010dcec708"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.864415 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-config" (OuterVolumeSpecName: "config") pod "fa2948a0-da02-4c95-bc2b-d9789564c9e7" (UID: "fa2948a0-da02-4c95-bc2b-d9789564c9e7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.868030 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fa2948a0-da02-4c95-bc2b-d9789564c9e7" (UID: "fa2948a0-da02-4c95-bc2b-d9789564c9e7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.870190 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fa2948a0-da02-4c95-bc2b-d9789564c9e7" (UID: "fa2948a0-da02-4c95-bc2b-d9789564c9e7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.871118 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:45:18 crc kubenswrapper[5035]: W1002 09:45:18.872377 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf42b1109_c1b4_4a08_a8d3_03cc50774a90.slice/crio-806d31a90204ebd1fcda3c8f699567c2cd0d3e8883ec2fabcdfeef0f31a68aee WatchSource:0}: Error finding container 806d31a90204ebd1fcda3c8f699567c2cd0d3e8883ec2fabcdfeef0f31a68aee: Status 404 returned error can't find the container with id 806d31a90204ebd1fcda3c8f699567c2cd0d3e8883ec2fabcdfeef0f31a68aee Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.891929 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.891963 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.891978 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dt5vs\" (UniqueName: \"kubernetes.io/projected/ce7048e6-3e78-4655-b89a-9f010dcec708-kube-api-access-dt5vs\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.891995 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892006 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95x2t\" (UniqueName: \"kubernetes.io/projected/fa2948a0-da02-4c95-bc2b-d9789564c9e7-kube-api-access-95x2t\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892016 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892024 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892031 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892069 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892079 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892086 5035 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ce7048e6-3e78-4655-b89a-9f010dcec708-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892094 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7048e6-3e78-4655-b89a-9f010dcec708-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.892102 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fa2948a0-da02-4c95-bc2b-d9789564c9e7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.909267 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 02 09:45:18 crc kubenswrapper[5035]: I1002 09:45:18.993803 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.119770 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.119761 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-h8tfl" event={"ID":"fa2948a0-da02-4c95-bc2b-d9789564c9e7","Type":"ContainerDied","Data":"db294ffa2d3eca59e8f6c713f2ff3b19ca901ec6e664878ab82bb6672aa44ee8"} Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.120544 5035 scope.go:117] "RemoveContainer" containerID="116c59c19ce9de2fa774d4767abea7fe77a099dde4c885273caa655d9818a6fc" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.127295 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ce7048e6-3e78-4655-b89a-9f010dcec708","Type":"ContainerDied","Data":"a51819184ed5687fabb590b4d418a1dd5038e7db762c760311f10d60dce35c37"} Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.127366 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.130390 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xb6lf" event={"ID":"2ad8a14e-58b5-4192-acf8-972ffdbd8e27","Type":"ContainerDied","Data":"17d5440a87165c8c0a691ee96c6243d513d632697ffd4c861e0d0f1490b718c2"} Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.130421 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17d5440a87165c8c0a691ee96c6243d513d632697ffd4c861e0d0f1490b718c2" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.130480 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xb6lf" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.132818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f42b1109-c1b4-4a08-a8d3-03cc50774a90","Type":"ContainerStarted","Data":"806d31a90204ebd1fcda3c8f699567c2cd0d3e8883ec2fabcdfeef0f31a68aee"} Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.134841 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-klp2h" event={"ID":"afad3fdc-5375-4bba-9832-f94381ba82aa","Type":"ContainerStarted","Data":"c27323c7e57c6556dc856f0f30e811673f692f7309ba9c7db35a02c8e13399ef"} Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.145293 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerStarted","Data":"1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7"} Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.147447 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-knpjm" event={"ID":"136b4da6-5c57-4a29-af5a-afe32c5f552f","Type":"ContainerDied","Data":"e78bfb5af27eaa5a80ca68cedcb1d6b7e1447fc366c25bae0e2944ad4d67d7e8"} Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.147484 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e78bfb5af27eaa5a80ca68cedcb1d6b7e1447fc366c25bae0e2944ad4d67d7e8" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.147508 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-knpjm" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.157156 5035 scope.go:117] "RemoveContainer" containerID="24dcf757f3fc909e46ee7b98193be6bd260d929e7cf0b4f62540c65634108952" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.172194 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-h8tfl"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.187785 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-h8tfl"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.219603 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.233186 5035 scope.go:117] "RemoveContainer" containerID="3ffd2a9e567ac65f44500bafeda0b9f77585a24cf6c1618dc1f27db173473226" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.236844 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.245427 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:45:19 crc kubenswrapper[5035]: E1002 09:45:19.245859 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerName="dnsmasq-dns" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.245876 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerName="dnsmasq-dns" Oct 02 09:45:19 crc kubenswrapper[5035]: E1002 09:45:19.245888 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-log" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.245895 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-log" Oct 02 09:45:19 crc kubenswrapper[5035]: E1002 09:45:19.245910 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="136b4da6-5c57-4a29-af5a-afe32c5f552f" containerName="placement-db-sync" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.245916 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="136b4da6-5c57-4a29-af5a-afe32c5f552f" containerName="placement-db-sync" Oct 02 09:45:19 crc kubenswrapper[5035]: E1002 09:45:19.245932 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad8a14e-58b5-4192-acf8-972ffdbd8e27" containerName="keystone-bootstrap" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.245938 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad8a14e-58b5-4192-acf8-972ffdbd8e27" containerName="keystone-bootstrap" Oct 02 09:45:19 crc kubenswrapper[5035]: E1002 09:45:19.245950 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-httpd" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.245958 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-httpd" Oct 02 09:45:19 crc kubenswrapper[5035]: E1002 09:45:19.245969 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerName="init" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.245975 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerName="init" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.246140 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-httpd" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.246160 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad8a14e-58b5-4192-acf8-972ffdbd8e27" containerName="keystone-bootstrap" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.246171 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" containerName="dnsmasq-dns" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.246180 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="136b4da6-5c57-4a29-af5a-afe32c5f552f" containerName="placement-db-sync" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.246189 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" containerName="glance-log" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.247105 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.251932 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.252250 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.256526 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.264121 5035 scope.go:117] "RemoveContainer" containerID="f6d79d95ef220440c49fd6178448d63000f20f5d772de27c03380b1dfe210aa3" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.405635 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.405683 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s9s2\" (UniqueName: \"kubernetes.io/projected/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-kube-api-access-5s9s2\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.405702 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-logs\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.405728 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.405767 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.405840 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.405944 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.406051 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.449548 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-59898c4b58-pz6vt"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.451283 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.456083 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.456145 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.456407 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.457804 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.458003 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fj9kx" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.458213 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.467762 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-59898c4b58-pz6vt"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507564 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507653 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507705 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507734 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507775 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507851 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507877 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s9s2\" (UniqueName: \"kubernetes.io/projected/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-kube-api-access-5s9s2\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.507896 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-logs\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.510407 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-logs\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.512082 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.512230 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.513684 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.521662 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.521788 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.530337 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.552665 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s9s2\" (UniqueName: \"kubernetes.io/projected/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-kube-api-access-5s9s2\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.554097 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-57bc66b65b-7wmtt"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.555612 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.559654 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.559857 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.559904 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-f7xgd" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.559865 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.560075 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.575544 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57bc66b65b-7wmtt"] Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.609789 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-scripts\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.609910 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-fernet-keys\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.610069 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-public-tls-certs\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.610131 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55kjt\" (UniqueName: \"kubernetes.io/projected/03747df9-2ce2-42b8-815a-92ffcce73253-kube-api-access-55kjt\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.610203 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-combined-ca-bundle\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.610238 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-internal-tls-certs\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.610299 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-config-data\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.610362 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-credential-keys\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.632714 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.712030 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55kjt\" (UniqueName: \"kubernetes.io/projected/03747df9-2ce2-42b8-815a-92ffcce73253-kube-api-access-55kjt\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.712923 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-combined-ca-bundle\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.712970 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzv68\" (UniqueName: \"kubernetes.io/projected/928cad66-0519-4d28-ae2f-a8f5875ac301-kube-api-access-tzv68\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.712989 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-internal-tls-certs\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713014 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-config-data\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713038 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-credential-keys\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713067 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928cad66-0519-4d28-ae2f-a8f5875ac301-logs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713092 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-combined-ca-bundle\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713137 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-public-tls-certs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713158 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-scripts\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713207 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-config-data\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713226 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-internal-tls-certs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713243 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-fernet-keys\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713279 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-scripts\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.713305 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-public-tls-certs\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.718837 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-internal-tls-certs\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.719363 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-public-tls-certs\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.722330 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-combined-ca-bundle\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.724219 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-credential-keys\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.725642 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-config-data\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.727020 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-scripts\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.727025 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/03747df9-2ce2-42b8-815a-92ffcce73253-fernet-keys\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.734104 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55kjt\" (UniqueName: \"kubernetes.io/projected/03747df9-2ce2-42b8-815a-92ffcce73253-kube-api-access-55kjt\") pod \"keystone-59898c4b58-pz6vt\" (UID: \"03747df9-2ce2-42b8-815a-92ffcce73253\") " pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.815604 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-scripts\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.815788 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzv68\" (UniqueName: \"kubernetes.io/projected/928cad66-0519-4d28-ae2f-a8f5875ac301-kube-api-access-tzv68\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.819896 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928cad66-0519-4d28-ae2f-a8f5875ac301-logs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.819969 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-combined-ca-bundle\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.820055 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-public-tls-certs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.820174 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-config-data\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.820223 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-internal-tls-certs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.820333 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/928cad66-0519-4d28-ae2f-a8f5875ac301-logs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.820441 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-scripts\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.823449 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-internal-tls-certs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.825246 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-public-tls-certs\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.825827 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-config-data\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.825973 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/928cad66-0519-4d28-ae2f-a8f5875ac301-combined-ca-bundle\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.837406 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzv68\" (UniqueName: \"kubernetes.io/projected/928cad66-0519-4d28-ae2f-a8f5875ac301-kube-api-access-tzv68\") pod \"placement-57bc66b65b-7wmtt\" (UID: \"928cad66-0519-4d28-ae2f-a8f5875ac301\") " pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.867732 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.931670 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:19 crc kubenswrapper[5035]: I1002 09:45:19.940460 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:20 crc kubenswrapper[5035]: I1002 09:45:20.208638 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce7048e6-3e78-4655-b89a-9f010dcec708" path="/var/lib/kubelet/pods/ce7048e6-3e78-4655-b89a-9f010dcec708/volumes" Oct 02 09:45:20 crc kubenswrapper[5035]: I1002 09:45:20.210655 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa2948a0-da02-4c95-bc2b-d9789564c9e7" path="/var/lib/kubelet/pods/fa2948a0-da02-4c95-bc2b-d9789564c9e7/volumes" Oct 02 09:45:20 crc kubenswrapper[5035]: I1002 09:45:20.211284 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f42b1109-c1b4-4a08-a8d3-03cc50774a90","Type":"ContainerStarted","Data":"c602845ee5438d0817f63a5900578511888dcf10b25a37336986d1f8931e2f29"} Oct 02 09:45:20 crc kubenswrapper[5035]: I1002 09:45:20.516795 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-59898c4b58-pz6vt"] Oct 02 09:45:20 crc kubenswrapper[5035]: I1002 09:45:20.671072 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57bc66b65b-7wmtt"] Oct 02 09:45:20 crc kubenswrapper[5035]: W1002 09:45:20.672111 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod928cad66_0519_4d28_ae2f_a8f5875ac301.slice/crio-2ec8340e0cefa8f6258477ed2a9f79742869844bcef6e883f344499a0ed16232 WatchSource:0}: Error finding container 2ec8340e0cefa8f6258477ed2a9f79742869844bcef6e883f344499a0ed16232: Status 404 returned error can't find the container with id 2ec8340e0cefa8f6258477ed2a9f79742869844bcef6e883f344499a0ed16232 Oct 02 09:45:20 crc kubenswrapper[5035]: I1002 09:45:20.691893 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.208525 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57bc66b65b-7wmtt" event={"ID":"928cad66-0519-4d28-ae2f-a8f5875ac301","Type":"ContainerStarted","Data":"47f9634c3e77624a0cc9642d6bb93eb24ff801d30d7b26b394fe8ff1ea9f33c5"} Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.208870 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57bc66b65b-7wmtt" event={"ID":"928cad66-0519-4d28-ae2f-a8f5875ac301","Type":"ContainerStarted","Data":"2ec8340e0cefa8f6258477ed2a9f79742869844bcef6e883f344499a0ed16232"} Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.210487 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442","Type":"ContainerStarted","Data":"0fcc82bf265838d8a9ec6e6d07cb82e193e3c0fceb522d410599cdf18c9f7ad8"} Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.213980 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-59898c4b58-pz6vt" event={"ID":"03747df9-2ce2-42b8-815a-92ffcce73253","Type":"ContainerStarted","Data":"01ed0d8ea712df2d8e4f03baec78c71bb3b4a7b785bd58c216a73827337580e7"} Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.214022 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-59898c4b58-pz6vt" event={"ID":"03747df9-2ce2-42b8-815a-92ffcce73253","Type":"ContainerStarted","Data":"fdc0cea48542888da5c988e1f044b6da88699e1a49ecba2edfea5497b68fd5c0"} Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.214055 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.218107 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f42b1109-c1b4-4a08-a8d3-03cc50774a90","Type":"ContainerStarted","Data":"804befe089b471e947159d43e917b186670d15949200f28417981b3bfffae808"} Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.230694 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-59898c4b58-pz6vt" podStartSLOduration=2.230677213 podStartE2EDuration="2.230677213s" podCreationTimestamp="2025-10-02 09:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:21.229396296 +0000 UTC m=+1086.585740341" watchObservedRunningTime="2025-10-02 09:45:21.230677213 +0000 UTC m=+1086.587021238" Oct 02 09:45:21 crc kubenswrapper[5035]: I1002 09:45:21.276224 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.276193869 podStartE2EDuration="8.276193869s" podCreationTimestamp="2025-10-02 09:45:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:21.272977186 +0000 UTC m=+1086.629321211" watchObservedRunningTime="2025-10-02 09:45:21.276193869 +0000 UTC m=+1086.632537894" Oct 02 09:45:22 crc kubenswrapper[5035]: I1002 09:45:22.234761 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57bc66b65b-7wmtt" event={"ID":"928cad66-0519-4d28-ae2f-a8f5875ac301","Type":"ContainerStarted","Data":"36c6b968408ccf6edf752deb67543b83a0783f4dd0bd4ba37eedd0a3e055b641"} Oct 02 09:45:22 crc kubenswrapper[5035]: I1002 09:45:22.235178 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:22 crc kubenswrapper[5035]: I1002 09:45:22.235198 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:22 crc kubenswrapper[5035]: I1002 09:45:22.238938 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442","Type":"ContainerStarted","Data":"f2c89a8d7b3ee8367edfb6af778e14bc3505cf972e0f8764ea0d989c7b3e5d34"} Oct 02 09:45:22 crc kubenswrapper[5035]: I1002 09:45:22.264294 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-57bc66b65b-7wmtt" podStartSLOduration=3.264275667 podStartE2EDuration="3.264275667s" podCreationTimestamp="2025-10-02 09:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:22.252247357 +0000 UTC m=+1087.608591392" watchObservedRunningTime="2025-10-02 09:45:22.264275667 +0000 UTC m=+1087.620619692" Oct 02 09:45:23 crc kubenswrapper[5035]: I1002 09:45:23.504724 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:23 crc kubenswrapper[5035]: I1002 09:45:23.505208 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:23 crc kubenswrapper[5035]: I1002 09:45:23.532913 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:23 crc kubenswrapper[5035]: I1002 09:45:23.545846 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:24 crc kubenswrapper[5035]: I1002 09:45:24.275631 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:24 crc kubenswrapper[5035]: I1002 09:45:24.275873 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:25 crc kubenswrapper[5035]: I1002 09:45:25.286570 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442","Type":"ContainerStarted","Data":"6ea426aa25725e5a82f3a3f4772279450441598ebdc4296ba755846b29b4fd6a"} Oct 02 09:45:25 crc kubenswrapper[5035]: I1002 09:45:25.320883 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.320856199 podStartE2EDuration="6.320856199s" podCreationTimestamp="2025-10-02 09:45:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:25.311337911 +0000 UTC m=+1090.667681956" watchObservedRunningTime="2025-10-02 09:45:25.320856199 +0000 UTC m=+1090.677200234" Oct 02 09:45:26 crc kubenswrapper[5035]: I1002 09:45:26.296310 5035 generic.go:334] "Generic (PLEG): container finished" podID="47413f27-08ae-470d-8877-84b1fd25fd8a" containerID="53ac0626fdd0ddba759654706b8203018a894f59965851289a2b8475b311e981" exitCode=0 Oct 02 09:45:26 crc kubenswrapper[5035]: I1002 09:45:26.296402 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d6j5s" event={"ID":"47413f27-08ae-470d-8877-84b1fd25fd8a","Type":"ContainerDied","Data":"53ac0626fdd0ddba759654706b8203018a894f59965851289a2b8475b311e981"} Oct 02 09:45:26 crc kubenswrapper[5035]: I1002 09:45:26.341594 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:27 crc kubenswrapper[5035]: I1002 09:45:27.253123 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.658038 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.720042 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7cd8\" (UniqueName: \"kubernetes.io/projected/47413f27-08ae-470d-8877-84b1fd25fd8a-kube-api-access-g7cd8\") pod \"47413f27-08ae-470d-8877-84b1fd25fd8a\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.720216 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-combined-ca-bundle\") pod \"47413f27-08ae-470d-8877-84b1fd25fd8a\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.720278 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-config\") pod \"47413f27-08ae-470d-8877-84b1fd25fd8a\" (UID: \"47413f27-08ae-470d-8877-84b1fd25fd8a\") " Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.736686 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47413f27-08ae-470d-8877-84b1fd25fd8a-kube-api-access-g7cd8" (OuterVolumeSpecName: "kube-api-access-g7cd8") pod "47413f27-08ae-470d-8877-84b1fd25fd8a" (UID: "47413f27-08ae-470d-8877-84b1fd25fd8a"). InnerVolumeSpecName "kube-api-access-g7cd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.747204 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47413f27-08ae-470d-8877-84b1fd25fd8a" (UID: "47413f27-08ae-470d-8877-84b1fd25fd8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.750725 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-config" (OuterVolumeSpecName: "config") pod "47413f27-08ae-470d-8877-84b1fd25fd8a" (UID: "47413f27-08ae-470d-8877-84b1fd25fd8a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.822219 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.822256 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/47413f27-08ae-470d-8877-84b1fd25fd8a-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:28 crc kubenswrapper[5035]: I1002 09:45:28.822267 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7cd8\" (UniqueName: \"kubernetes.io/projected/47413f27-08ae-470d-8877-84b1fd25fd8a-kube-api-access-g7cd8\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.330387 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-d6j5s" event={"ID":"47413f27-08ae-470d-8877-84b1fd25fd8a","Type":"ContainerDied","Data":"3acf39cbd430ffbcf8dfc9049422d0dbbd5414a7654ee36150ec6c8dfeadf78b"} Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.330430 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3acf39cbd430ffbcf8dfc9049422d0dbbd5414a7654ee36150ec6c8dfeadf78b" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.330993 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-d6j5s" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.868383 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.868742 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.939017 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tv4h4"] Oct 02 09:45:29 crc kubenswrapper[5035]: E1002 09:45:29.939423 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47413f27-08ae-470d-8877-84b1fd25fd8a" containerName="neutron-db-sync" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.939440 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="47413f27-08ae-470d-8877-84b1fd25fd8a" containerName="neutron-db-sync" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.939643 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="47413f27-08ae-470d-8877-84b1fd25fd8a" containerName="neutron-db-sync" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.940576 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.953813 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.953900 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 09:45:29 crc kubenswrapper[5035]: I1002 09:45:29.964963 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tv4h4"] Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.044374 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.044437 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-config\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.044462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.044513 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.044580 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjctt\" (UniqueName: \"kubernetes.io/projected/d283583a-e964-4817-b474-8c128a4e78b6-kube-api-access-qjctt\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.044605 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.054570 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-58f6c5cff8-5t92g"] Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.056115 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.058658 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.058794 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-46h8t" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.058911 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.059002 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.071695 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58f6c5cff8-5t92g"] Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146269 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146332 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-config\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146399 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-combined-ca-bundle\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146419 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146445 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-httpd-config\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146496 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjctt\" (UniqueName: \"kubernetes.io/projected/d283583a-e964-4817-b474-8c128a4e78b6-kube-api-access-qjctt\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146520 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146573 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-config\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146586 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-ovndb-tls-certs\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.146610 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz72q\" (UniqueName: \"kubernetes.io/projected/ee399cdc-62d4-409d-8661-5ff797758647-kube-api-access-lz72q\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.147468 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.148097 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-config\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.148754 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.148779 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.148794 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.166521 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjctt\" (UniqueName: \"kubernetes.io/projected/d283583a-e964-4817-b474-8c128a4e78b6-kube-api-access-qjctt\") pod \"dnsmasq-dns-84b966f6c9-tv4h4\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.249959 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz72q\" (UniqueName: \"kubernetes.io/projected/ee399cdc-62d4-409d-8661-5ff797758647-kube-api-access-lz72q\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.250478 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-combined-ca-bundle\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.250515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-httpd-config\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.250742 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-config\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.250781 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-ovndb-tls-certs\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.254415 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-httpd-config\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.264225 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-combined-ca-bundle\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.265001 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-config\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.265586 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-ovndb-tls-certs\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.266785 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.268700 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz72q\" (UniqueName: \"kubernetes.io/projected/ee399cdc-62d4-409d-8661-5ff797758647-kube-api-access-lz72q\") pod \"neutron-58f6c5cff8-5t92g\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.340720 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.340781 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.377318 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:30 crc kubenswrapper[5035]: I1002 09:45:30.877980 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tv4h4"] Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.087495 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-58f6c5cff8-5t92g"] Oct 02 09:45:31 crc kubenswrapper[5035]: E1002 09:45:31.320429 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd283583a_e964_4817_b474_8c128a4e78b6.slice/crio-conmon-1936d1cec2a4726fe036f2f0c7c249ac4ae357d427f4bfe4d9242a22e0c8f2a0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd283583a_e964_4817_b474_8c128a4e78b6.slice/crio-1936d1cec2a4726fe036f2f0c7c249ac4ae357d427f4bfe4d9242a22e0c8f2a0.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.356556 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qf6x5" event={"ID":"830b460b-be63-4b42-a442-80d844ef6908","Type":"ContainerStarted","Data":"6a4f10c5814a8ae7e34611ca2236badbceaa7dd1b30aa7238bd4b1a43e72fa80"} Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.363581 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerStarted","Data":"5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7"} Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.363653 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-central-agent" containerID="cri-o://17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8" gracePeriod=30 Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.363731 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="sg-core" containerID="cri-o://1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7" gracePeriod=30 Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.363747 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="proxy-httpd" containerID="cri-o://5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7" gracePeriod=30 Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.363767 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.363852 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-notification-agent" containerID="cri-o://04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673" gracePeriod=30 Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.373370 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7t2tz" event={"ID":"76ce8ea1-cda3-4f68-9570-c6c02b56283f","Type":"ContainerStarted","Data":"49e6bc1dcd3210e3ae2c9a69fafc0c053a9f33d4a050ddbc3b8504564fdd8373"} Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.375970 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-qf6x5" podStartSLOduration=3.160153528 podStartE2EDuration="52.375952901s" podCreationTimestamp="2025-10-02 09:44:39 +0000 UTC" firstStartedPulling="2025-10-02 09:44:41.152321136 +0000 UTC m=+1046.508665161" lastFinishedPulling="2025-10-02 09:45:30.368120509 +0000 UTC m=+1095.724464534" observedRunningTime="2025-10-02 09:45:31.373718535 +0000 UTC m=+1096.730062560" watchObservedRunningTime="2025-10-02 09:45:31.375952901 +0000 UTC m=+1096.732296926" Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.393326 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7t2tz" podStartSLOduration=3.12038191 podStartE2EDuration="52.393304566s" podCreationTimestamp="2025-10-02 09:44:39 +0000 UTC" firstStartedPulling="2025-10-02 09:44:41.103290518 +0000 UTC m=+1046.459634543" lastFinishedPulling="2025-10-02 09:45:30.376213164 +0000 UTC m=+1095.732557199" observedRunningTime="2025-10-02 09:45:31.388848186 +0000 UTC m=+1096.745192211" watchObservedRunningTime="2025-10-02 09:45:31.393304566 +0000 UTC m=+1096.749648591" Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.397198 5035 generic.go:334] "Generic (PLEG): container finished" podID="afad3fdc-5375-4bba-9832-f94381ba82aa" containerID="d4c76bc0185947a80f0b2410ab59770810989ba1b07a74f4c6a32b8216dd60f0" exitCode=0 Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.397272 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-klp2h" event={"ID":"afad3fdc-5375-4bba-9832-f94381ba82aa","Type":"ContainerDied","Data":"d4c76bc0185947a80f0b2410ab59770810989ba1b07a74f4c6a32b8216dd60f0"} Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.404468 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58f6c5cff8-5t92g" event={"ID":"ee399cdc-62d4-409d-8661-5ff797758647","Type":"ContainerStarted","Data":"4acdd4bc747f8147c13663bc271cd6dbddbe356e9da7759ff2087a83e5540ec0"} Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.413618 5035 generic.go:334] "Generic (PLEG): container finished" podID="d283583a-e964-4817-b474-8c128a4e78b6" containerID="1936d1cec2a4726fe036f2f0c7c249ac4ae357d427f4bfe4d9242a22e0c8f2a0" exitCode=0 Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.415283 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" event={"ID":"d283583a-e964-4817-b474-8c128a4e78b6","Type":"ContainerDied","Data":"1936d1cec2a4726fe036f2f0c7c249ac4ae357d427f4bfe4d9242a22e0c8f2a0"} Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.415322 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" event={"ID":"d283583a-e964-4817-b474-8c128a4e78b6","Type":"ContainerStarted","Data":"31593d12b0ea1d889fcb7cfa4261d6b089b130a719d46841fb0782ca6aaa2c64"} Oct 02 09:45:31 crc kubenswrapper[5035]: I1002 09:45:31.435908 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.312064673 podStartE2EDuration="52.435880306s" podCreationTimestamp="2025-10-02 09:44:39 +0000 UTC" firstStartedPulling="2025-10-02 09:44:41.245620904 +0000 UTC m=+1046.601964929" lastFinishedPulling="2025-10-02 09:45:30.369436537 +0000 UTC m=+1095.725780562" observedRunningTime="2025-10-02 09:45:31.419984933 +0000 UTC m=+1096.776328968" watchObservedRunningTime="2025-10-02 09:45:31.435880306 +0000 UTC m=+1096.792224331" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.089425 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6f9c95b6cc-5grzm"] Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.090861 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.093128 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.093238 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.111306 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f9c95b6cc-5grzm"] Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.195431 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-combined-ca-bundle\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.195792 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-config\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.195868 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-internal-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.196682 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-public-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.196778 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-ovndb-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.196848 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-httpd-config\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.196882 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsmsb\" (UniqueName: \"kubernetes.io/projected/6cfcc6bb-1b26-42a1-812e-96bd03767545-kube-api-access-tsmsb\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.298672 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-ovndb-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.298746 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-httpd-config\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.298787 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsmsb\" (UniqueName: \"kubernetes.io/projected/6cfcc6bb-1b26-42a1-812e-96bd03767545-kube-api-access-tsmsb\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.298876 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-combined-ca-bundle\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.298910 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-config\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.298977 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-internal-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.299026 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-public-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.302991 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-httpd-config\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.303835 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-combined-ca-bundle\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.305263 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-internal-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.305742 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-config\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.306367 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-ovndb-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.313567 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfcc6bb-1b26-42a1-812e-96bd03767545-public-tls-certs\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.321061 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsmsb\" (UniqueName: \"kubernetes.io/projected/6cfcc6bb-1b26-42a1-812e-96bd03767545-kube-api-access-tsmsb\") pod \"neutron-6f9c95b6cc-5grzm\" (UID: \"6cfcc6bb-1b26-42a1-812e-96bd03767545\") " pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.414502 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.434051 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-klp2h" event={"ID":"afad3fdc-5375-4bba-9832-f94381ba82aa","Type":"ContainerStarted","Data":"9ed1c71145077c8e62f1a887d9e9e42bc8652299ccafecb1821b2fdf53bc90d5"} Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.447368 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58f6c5cff8-5t92g" event={"ID":"ee399cdc-62d4-409d-8661-5ff797758647","Type":"ContainerStarted","Data":"173049a936e4990ec816ab9bbdb1fbeca88ec082a9aa7593cdea4d1076f4c304"} Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.447462 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58f6c5cff8-5t92g" event={"ID":"ee399cdc-62d4-409d-8661-5ff797758647","Type":"ContainerStarted","Data":"95640f4b076c18d15a63d7cfac95605fe0fce07eb6470486af1f04e53256acac"} Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.448838 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.457847 5035 generic.go:334] "Generic (PLEG): container finished" podID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerID="5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7" exitCode=0 Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.457894 5035 generic.go:334] "Generic (PLEG): container finished" podID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerID="1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7" exitCode=2 Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.457905 5035 generic.go:334] "Generic (PLEG): container finished" podID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerID="17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8" exitCode=0 Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.457980 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerDied","Data":"5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7"} Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.458012 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerDied","Data":"1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7"} Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.458024 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerDied","Data":"17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8"} Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.475806 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" event={"ID":"d283583a-e964-4817-b474-8c128a4e78b6","Type":"ContainerStarted","Data":"27503ecef9b6e6b735d233e00872ea1d2f8da36bbb55086aa5412dbd44a31c95"} Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.479821 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.488184 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-db-sync-klp2h" podStartSLOduration=6.039618769 podStartE2EDuration="17.488161854s" podCreationTimestamp="2025-10-02 09:45:15 +0000 UTC" firstStartedPulling="2025-10-02 09:45:18.784932982 +0000 UTC m=+1084.141277007" lastFinishedPulling="2025-10-02 09:45:30.233476067 +0000 UTC m=+1095.589820092" observedRunningTime="2025-10-02 09:45:32.458467469 +0000 UTC m=+1097.814811504" watchObservedRunningTime="2025-10-02 09:45:32.488161854 +0000 UTC m=+1097.844505889" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.491662 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-58f6c5cff8-5t92g" podStartSLOduration=2.491641785 podStartE2EDuration="2.491641785s" podCreationTimestamp="2025-10-02 09:45:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:32.477876044 +0000 UTC m=+1097.834220069" watchObservedRunningTime="2025-10-02 09:45:32.491641785 +0000 UTC m=+1097.847985810" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.528386 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" podStartSLOduration=3.528361384 podStartE2EDuration="3.528361384s" podCreationTimestamp="2025-10-02 09:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:32.516405656 +0000 UTC m=+1097.872749681" watchObservedRunningTime="2025-10-02 09:45:32.528361384 +0000 UTC m=+1097.884705409" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.767358 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.767503 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.768048 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6f9c95b6cc-5grzm"] Oct 02 09:45:32 crc kubenswrapper[5035]: I1002 09:45:32.768151 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 09:45:32 crc kubenswrapper[5035]: W1002 09:45:32.769653 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cfcc6bb_1b26_42a1_812e_96bd03767545.slice/crio-6e1b38e370b5c4f3a3b295680001d9f067998be12d8757ff8d7087c660b1e7d3 WatchSource:0}: Error finding container 6e1b38e370b5c4f3a3b295680001d9f067998be12d8757ff8d7087c660b1e7d3: Status 404 returned error can't find the container with id 6e1b38e370b5c4f3a3b295680001d9f067998be12d8757ff8d7087c660b1e7d3 Oct 02 09:45:33 crc kubenswrapper[5035]: I1002 09:45:33.485341 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f9c95b6cc-5grzm" event={"ID":"6cfcc6bb-1b26-42a1-812e-96bd03767545","Type":"ContainerStarted","Data":"b37615b9e5aa6e2be3f97ed5b676e74dd924bd5431fce1b92bcc7d58162c5e9b"} Oct 02 09:45:33 crc kubenswrapper[5035]: I1002 09:45:33.485699 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f9c95b6cc-5grzm" event={"ID":"6cfcc6bb-1b26-42a1-812e-96bd03767545","Type":"ContainerStarted","Data":"70ed0c73fb91e9a2b22539c7c48c40ac810767db1efb7ce145c9c5ae859984c0"} Oct 02 09:45:33 crc kubenswrapper[5035]: I1002 09:45:33.485714 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6f9c95b6cc-5grzm" event={"ID":"6cfcc6bb-1b26-42a1-812e-96bd03767545","Type":"ContainerStarted","Data":"6e1b38e370b5c4f3a3b295680001d9f067998be12d8757ff8d7087c660b1e7d3"} Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.192122 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.246859 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-log-httpd\") pod \"3636a749-409c-4b58-a29f-8ca1568d2dbc\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.246960 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-config-data\") pod \"3636a749-409c-4b58-a29f-8ca1568d2dbc\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.247036 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r88rm\" (UniqueName: \"kubernetes.io/projected/3636a749-409c-4b58-a29f-8ca1568d2dbc-kube-api-access-r88rm\") pod \"3636a749-409c-4b58-a29f-8ca1568d2dbc\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.247080 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-sg-core-conf-yaml\") pod \"3636a749-409c-4b58-a29f-8ca1568d2dbc\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.247124 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-scripts\") pod \"3636a749-409c-4b58-a29f-8ca1568d2dbc\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.247178 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-combined-ca-bundle\") pod \"3636a749-409c-4b58-a29f-8ca1568d2dbc\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.247208 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-run-httpd\") pod \"3636a749-409c-4b58-a29f-8ca1568d2dbc\" (UID: \"3636a749-409c-4b58-a29f-8ca1568d2dbc\") " Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.247563 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3636a749-409c-4b58-a29f-8ca1568d2dbc" (UID: "3636a749-409c-4b58-a29f-8ca1568d2dbc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.251935 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3636a749-409c-4b58-a29f-8ca1568d2dbc" (UID: "3636a749-409c-4b58-a29f-8ca1568d2dbc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.266752 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3636a749-409c-4b58-a29f-8ca1568d2dbc-kube-api-access-r88rm" (OuterVolumeSpecName: "kube-api-access-r88rm") pod "3636a749-409c-4b58-a29f-8ca1568d2dbc" (UID: "3636a749-409c-4b58-a29f-8ca1568d2dbc"). InnerVolumeSpecName "kube-api-access-r88rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.266952 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-scripts" (OuterVolumeSpecName: "scripts") pod "3636a749-409c-4b58-a29f-8ca1568d2dbc" (UID: "3636a749-409c-4b58-a29f-8ca1568d2dbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.327150 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6f9c95b6cc-5grzm" podStartSLOduration=2.327132352 podStartE2EDuration="2.327132352s" podCreationTimestamp="2025-10-02 09:45:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:33.516563326 +0000 UTC m=+1098.872907361" watchObservedRunningTime="2025-10-02 09:45:34.327132352 +0000 UTC m=+1099.683476377" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.330719 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3636a749-409c-4b58-a29f-8ca1568d2dbc" (UID: "3636a749-409c-4b58-a29f-8ca1568d2dbc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.350493 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r88rm\" (UniqueName: \"kubernetes.io/projected/3636a749-409c-4b58-a29f-8ca1568d2dbc-kube-api-access-r88rm\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.350548 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.350558 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.350566 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.350576 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3636a749-409c-4b58-a29f-8ca1568d2dbc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.395265 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3636a749-409c-4b58-a29f-8ca1568d2dbc" (UID: "3636a749-409c-4b58-a29f-8ca1568d2dbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.451484 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.519680 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-config-data" (OuterVolumeSpecName: "config-data") pod "3636a749-409c-4b58-a29f-8ca1568d2dbc" (UID: "3636a749-409c-4b58-a29f-8ca1568d2dbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.521755 5035 generic.go:334] "Generic (PLEG): container finished" podID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerID="04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673" exitCode=0 Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.522658 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.522684 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerDied","Data":"04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673"} Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.522709 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3636a749-409c-4b58-a29f-8ca1568d2dbc","Type":"ContainerDied","Data":"ff97859ea3e14f77f87fda3eb564364b6d16d9caef16589bb892ca19cc9e882a"} Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.522729 5035 scope.go:117] "RemoveContainer" containerID="5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.523842 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.547717 5035 scope.go:117] "RemoveContainer" containerID="1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.561143 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3636a749-409c-4b58-a29f-8ca1568d2dbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.573380 5035 scope.go:117] "RemoveContainer" containerID="04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.594033 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.608744 5035 scope.go:117] "RemoveContainer" containerID="17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.616519 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.628640 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.629174 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="proxy-httpd" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629200 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="proxy-httpd" Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.629229 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-notification-agent" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629238 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-notification-agent" Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.629252 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="sg-core" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629263 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="sg-core" Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.629295 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-central-agent" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629303 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-central-agent" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629574 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="sg-core" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629601 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="proxy-httpd" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629615 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-central-agent" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.629632 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" containerName="ceilometer-notification-agent" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.631787 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.636617 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.636799 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.637695 5035 scope.go:117] "RemoveContainer" containerID="5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.637844 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.641931 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7\": container with ID starting with 5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7 not found: ID does not exist" containerID="5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.641977 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7"} err="failed to get container status \"5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7\": rpc error: code = NotFound desc = could not find container \"5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7\": container with ID starting with 5242c400e5983d5cb46224a2367dccd1d66d137988ff692c33bf38cd6a25c8a7 not found: ID does not exist" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.642009 5035 scope.go:117] "RemoveContainer" containerID="1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7" Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.642491 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7\": container with ID starting with 1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7 not found: ID does not exist" containerID="1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.642521 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7"} err="failed to get container status \"1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7\": rpc error: code = NotFound desc = could not find container \"1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7\": container with ID starting with 1463784c1e12d5e6ea05fd151e8c8412d63135e39cb2db5ae4800748ea0312a7 not found: ID does not exist" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.642552 5035 scope.go:117] "RemoveContainer" containerID="04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673" Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.642821 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673\": container with ID starting with 04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673 not found: ID does not exist" containerID="04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.644753 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673"} err="failed to get container status \"04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673\": rpc error: code = NotFound desc = could not find container \"04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673\": container with ID starting with 04c112414d308e6dcc10e8b1626f923f7fa533e53cebad6e8f8a5a446e27a673 not found: ID does not exist" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.644839 5035 scope.go:117] "RemoveContainer" containerID="17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8" Oct 02 09:45:34 crc kubenswrapper[5035]: E1002 09:45:34.648739 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8\": container with ID starting with 17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8 not found: ID does not exist" containerID="17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.648805 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8"} err="failed to get container status \"17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8\": rpc error: code = NotFound desc = could not find container \"17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8\": container with ID starting with 17aa8db6cfd13fac824538645f9df883bf5b5c9d1c4ff10b90599adc5ebc87f8 not found: ID does not exist" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.763799 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-scripts\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.763860 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.763890 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mct2k\" (UniqueName: \"kubernetes.io/projected/1275078f-89d8-4873-9ae2-4be7336e399b-kube-api-access-mct2k\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.763934 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-log-httpd\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.764001 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-run-httpd\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.764154 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-config-data\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.764232 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.865798 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-scripts\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.865860 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.865896 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mct2k\" (UniqueName: \"kubernetes.io/projected/1275078f-89d8-4873-9ae2-4be7336e399b-kube-api-access-mct2k\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.865968 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-log-httpd\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.865999 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-run-httpd\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.866055 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-config-data\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.866093 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.866579 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-run-httpd\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.866583 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-log-httpd\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.873327 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.875000 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-scripts\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.879129 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.885025 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-config-data\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.887421 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mct2k\" (UniqueName: \"kubernetes.io/projected/1275078f-89d8-4873-9ae2-4be7336e399b-kube-api-access-mct2k\") pod \"ceilometer-0\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " pod="openstack/ceilometer-0" Oct 02 09:45:34 crc kubenswrapper[5035]: I1002 09:45:34.971160 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:45:35 crc kubenswrapper[5035]: I1002 09:45:35.448349 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:45:35 crc kubenswrapper[5035]: I1002 09:45:35.531059 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerStarted","Data":"5808dffd90cdd0c2dbbda94dda8993709f62e9f250345547b8b486b813b949ad"} Oct 02 09:45:35 crc kubenswrapper[5035]: I1002 09:45:35.533618 5035 generic.go:334] "Generic (PLEG): container finished" podID="830b460b-be63-4b42-a442-80d844ef6908" containerID="6a4f10c5814a8ae7e34611ca2236badbceaa7dd1b30aa7238bd4b1a43e72fa80" exitCode=0 Oct 02 09:45:35 crc kubenswrapper[5035]: I1002 09:45:35.533679 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qf6x5" event={"ID":"830b460b-be63-4b42-a442-80d844ef6908","Type":"ContainerDied","Data":"6a4f10c5814a8ae7e34611ca2236badbceaa7dd1b30aa7238bd4b1a43e72fa80"} Oct 02 09:45:36 crc kubenswrapper[5035]: I1002 09:45:36.175283 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3636a749-409c-4b58-a29f-8ca1568d2dbc" path="/var/lib/kubelet/pods/3636a749-409c-4b58-a29f-8ca1568d2dbc/volumes" Oct 02 09:45:36 crc kubenswrapper[5035]: I1002 09:45:36.547044 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerStarted","Data":"df35b8f6c564cd8b61a2c0019a9f5f3ef940a2300bd8cc70ac9eeff6bb4288f0"} Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.004589 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.010120 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-db-sync-config-data\") pod \"830b460b-be63-4b42-a442-80d844ef6908\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.010155 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-combined-ca-bundle\") pod \"830b460b-be63-4b42-a442-80d844ef6908\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.010215 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9rvk\" (UniqueName: \"kubernetes.io/projected/830b460b-be63-4b42-a442-80d844ef6908-kube-api-access-f9rvk\") pod \"830b460b-be63-4b42-a442-80d844ef6908\" (UID: \"830b460b-be63-4b42-a442-80d844ef6908\") " Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.016290 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/830b460b-be63-4b42-a442-80d844ef6908-kube-api-access-f9rvk" (OuterVolumeSpecName: "kube-api-access-f9rvk") pod "830b460b-be63-4b42-a442-80d844ef6908" (UID: "830b460b-be63-4b42-a442-80d844ef6908"). InnerVolumeSpecName "kube-api-access-f9rvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.028639 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "830b460b-be63-4b42-a442-80d844ef6908" (UID: "830b460b-be63-4b42-a442-80d844ef6908"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.092642 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "830b460b-be63-4b42-a442-80d844ef6908" (UID: "830b460b-be63-4b42-a442-80d844ef6908"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.111315 5035 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.111344 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/830b460b-be63-4b42-a442-80d844ef6908-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.111354 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9rvk\" (UniqueName: \"kubernetes.io/projected/830b460b-be63-4b42-a442-80d844ef6908-kube-api-access-f9rvk\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.557586 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerStarted","Data":"2d07daec3490e305d5ebf660b7f949167416a0421bca4ec510dd7ab6235adb3e"} Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.560196 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-qf6x5" event={"ID":"830b460b-be63-4b42-a442-80d844ef6908","Type":"ContainerDied","Data":"41e6d8f29f456dd758a6eefe2b97cd495b0d9fe972266a920c8379ff09ea006d"} Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.560244 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41e6d8f29f456dd758a6eefe2b97cd495b0d9fe972266a920c8379ff09ea006d" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.560311 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-qf6x5" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.859354 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-545df9b758-drjst"] Oct 02 09:45:37 crc kubenswrapper[5035]: E1002 09:45:37.861770 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="830b460b-be63-4b42-a442-80d844ef6908" containerName="barbican-db-sync" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.861797 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="830b460b-be63-4b42-a442-80d844ef6908" containerName="barbican-db-sync" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.862037 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="830b460b-be63-4b42-a442-80d844ef6908" containerName="barbican-db-sync" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.864444 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.869873 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-phxhr" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.870051 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.870131 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.904048 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5d9b9799bc-tvwd9"] Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.905822 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.922009 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936034 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-config-data\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936078 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54742b9b-0346-4cdb-9bf2-75f35c59beaf-logs\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936095 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-combined-ca-bundle\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936114 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbrvd\" (UniqueName: \"kubernetes.io/projected/76945d28-7017-4b37-ac66-c356e9190230-kube-api-access-pbrvd\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936146 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-config-data\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936181 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-config-data-custom\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936199 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-combined-ca-bundle\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936221 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-config-data-custom\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936282 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76945d28-7017-4b37-ac66-c356e9190230-logs\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.936297 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n46qr\" (UniqueName: \"kubernetes.io/projected/54742b9b-0346-4cdb-9bf2-75f35c59beaf-kube-api-access-n46qr\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.938787 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-545df9b758-drjst"] Oct 02 09:45:37 crc kubenswrapper[5035]: I1002 09:45:37.960300 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d9b9799bc-tvwd9"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.015668 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tv4h4"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.015927 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" podUID="d283583a-e964-4817-b474-8c128a4e78b6" containerName="dnsmasq-dns" containerID="cri-o://27503ecef9b6e6b735d233e00872ea1d2f8da36bbb55086aa5412dbd44a31c95" gracePeriod=10 Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.017981 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.035192 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-ns7jc"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.036821 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037691 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-combined-ca-bundle\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037729 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbrvd\" (UniqueName: \"kubernetes.io/projected/76945d28-7017-4b37-ac66-c356e9190230-kube-api-access-pbrvd\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037766 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-config-data\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037806 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037827 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-config-data-custom\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037849 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037868 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-combined-ca-bundle\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037893 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-config-data-custom\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037933 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037953 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-config\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.037993 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n46qr\" (UniqueName: \"kubernetes.io/projected/54742b9b-0346-4cdb-9bf2-75f35c59beaf-kube-api-access-n46qr\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.038014 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76945d28-7017-4b37-ac66-c356e9190230-logs\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.038046 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqczm\" (UniqueName: \"kubernetes.io/projected/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-kube-api-access-hqczm\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.038071 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-config-data\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.038093 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.038114 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54742b9b-0346-4cdb-9bf2-75f35c59beaf-logs\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.038444 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54742b9b-0346-4cdb-9bf2-75f35c59beaf-logs\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.039934 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76945d28-7017-4b37-ac66-c356e9190230-logs\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.059224 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-f6c775994-6t2fx"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.061119 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.070908 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-config-data\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.071082 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.071714 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-config-data-custom\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.076508 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-combined-ca-bundle\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.076786 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-config-data\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.077504 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54742b9b-0346-4cdb-9bf2-75f35c59beaf-config-data-custom\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.077719 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n46qr\" (UniqueName: \"kubernetes.io/projected/54742b9b-0346-4cdb-9bf2-75f35c59beaf-kube-api-access-n46qr\") pod \"barbican-keystone-listener-545df9b758-drjst\" (UID: \"54742b9b-0346-4cdb-9bf2-75f35c59beaf\") " pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.078648 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76945d28-7017-4b37-ac66-c356e9190230-combined-ca-bundle\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.081459 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbrvd\" (UniqueName: \"kubernetes.io/projected/76945d28-7017-4b37-ac66-c356e9190230-kube-api-access-pbrvd\") pod \"barbican-worker-5d9b9799bc-tvwd9\" (UID: \"76945d28-7017-4b37-ac66-c356e9190230\") " pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.121594 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-ns7jc"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.140895 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.141007 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-config\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.141089 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqczm\" (UniqueName: \"kubernetes.io/projected/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-kube-api-access-hqczm\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.141144 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.141258 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.141294 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.144111 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.144259 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.144583 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-config\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.147429 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f6c775994-6t2fx"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.148095 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.148171 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.162166 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqczm\" (UniqueName: \"kubernetes.io/projected/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-kube-api-access-hqczm\") pod \"dnsmasq-dns-75c8ddd69c-ns7jc\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.199555 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5d9b9799bc-tvwd9" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.233101 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.244829 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.244928 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data-custom\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.245040 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91467c75-7c7d-4b20-baad-b168b8eabcbb-logs\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.245084 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-combined-ca-bundle\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.245223 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9qzg\" (UniqueName: \"kubernetes.io/projected/91467c75-7c7d-4b20-baad-b168b8eabcbb-kube-api-access-d9qzg\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.340026 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-545df9b758-drjst" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.347106 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91467c75-7c7d-4b20-baad-b168b8eabcbb-logs\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.347157 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-combined-ca-bundle\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.347260 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9qzg\" (UniqueName: \"kubernetes.io/projected/91467c75-7c7d-4b20-baad-b168b8eabcbb-kube-api-access-d9qzg\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.347320 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.347358 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data-custom\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.353031 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91467c75-7c7d-4b20-baad-b168b8eabcbb-logs\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.354922 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-combined-ca-bundle\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.358104 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data-custom\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.359102 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.368410 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9qzg\" (UniqueName: \"kubernetes.io/projected/91467c75-7c7d-4b20-baad-b168b8eabcbb-kube-api-access-d9qzg\") pod \"barbican-api-f6c775994-6t2fx\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.568055 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.591463 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerStarted","Data":"7e2460126f8043a6ac993c056ac4f1abd39edf64d7d97920a5d162c9ebb0c738"} Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.594080 5035 generic.go:334] "Generic (PLEG): container finished" podID="d283583a-e964-4817-b474-8c128a4e78b6" containerID="27503ecef9b6e6b735d233e00872ea1d2f8da36bbb55086aa5412dbd44a31c95" exitCode=0 Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.594111 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" event={"ID":"d283583a-e964-4817-b474-8c128a4e78b6","Type":"ContainerDied","Data":"27503ecef9b6e6b735d233e00872ea1d2f8da36bbb55086aa5412dbd44a31c95"} Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.656883 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.757862 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-svc\") pod \"d283583a-e964-4817-b474-8c128a4e78b6\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.758079 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-nb\") pod \"d283583a-e964-4817-b474-8c128a4e78b6\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.758816 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjctt\" (UniqueName: \"kubernetes.io/projected/d283583a-e964-4817-b474-8c128a4e78b6-kube-api-access-qjctt\") pod \"d283583a-e964-4817-b474-8c128a4e78b6\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.758868 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-config\") pod \"d283583a-e964-4817-b474-8c128a4e78b6\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.758967 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-sb\") pod \"d283583a-e964-4817-b474-8c128a4e78b6\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.759011 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-swift-storage-0\") pod \"d283583a-e964-4817-b474-8c128a4e78b6\" (UID: \"d283583a-e964-4817-b474-8c128a4e78b6\") " Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.765393 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d283583a-e964-4817-b474-8c128a4e78b6-kube-api-access-qjctt" (OuterVolumeSpecName: "kube-api-access-qjctt") pod "d283583a-e964-4817-b474-8c128a4e78b6" (UID: "d283583a-e964-4817-b474-8c128a4e78b6"). InnerVolumeSpecName "kube-api-access-qjctt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.829786 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d283583a-e964-4817-b474-8c128a4e78b6" (UID: "d283583a-e964-4817-b474-8c128a4e78b6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.835726 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5d9b9799bc-tvwd9"] Oct 02 09:45:38 crc kubenswrapper[5035]: W1002 09:45:38.840750 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76945d28_7017_4b37_ac66_c356e9190230.slice/crio-17703fd86227714ae35dcfd5c08c07fcf868c83008952e0bb1fde369d0d65166 WatchSource:0}: Error finding container 17703fd86227714ae35dcfd5c08c07fcf868c83008952e0bb1fde369d0d65166: Status 404 returned error can't find the container with id 17703fd86227714ae35dcfd5c08c07fcf868c83008952e0bb1fde369d0d65166 Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.844091 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d283583a-e964-4817-b474-8c128a4e78b6" (UID: "d283583a-e964-4817-b474-8c128a4e78b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.853265 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d283583a-e964-4817-b474-8c128a4e78b6" (UID: "d283583a-e964-4817-b474-8c128a4e78b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.861187 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.861220 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjctt\" (UniqueName: \"kubernetes.io/projected/d283583a-e964-4817-b474-8c128a4e78b6-kube-api-access-qjctt\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.861231 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.861239 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.872798 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-config" (OuterVolumeSpecName: "config") pod "d283583a-e964-4817-b474-8c128a4e78b6" (UID: "d283583a-e964-4817-b474-8c128a4e78b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.878705 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-ns7jc"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.894653 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d283583a-e964-4817-b474-8c128a4e78b6" (UID: "d283583a-e964-4817-b474-8c128a4e78b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.956209 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-545df9b758-drjst"] Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.963350 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:38 crc kubenswrapper[5035]: I1002 09:45:38.963394 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d283583a-e964-4817-b474-8c128a4e78b6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.142545 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-f6c775994-6t2fx"] Oct 02 09:45:39 crc kubenswrapper[5035]: W1002 09:45:39.149294 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91467c75_7c7d_4b20_baad_b168b8eabcbb.slice/crio-d969915ec6df8b8dbe9c2a65574f09d24938a93579fa72956de4f1f392ca1778 WatchSource:0}: Error finding container d969915ec6df8b8dbe9c2a65574f09d24938a93579fa72956de4f1f392ca1778: Status 404 returned error can't find the container with id d969915ec6df8b8dbe9c2a65574f09d24938a93579fa72956de4f1f392ca1778 Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.618594 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6c775994-6t2fx" event={"ID":"91467c75-7c7d-4b20-baad-b168b8eabcbb","Type":"ContainerStarted","Data":"d93282a16895c929b93ba1290144e9be905106614639e0cfc8df3cb492d7061e"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.618640 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6c775994-6t2fx" event={"ID":"91467c75-7c7d-4b20-baad-b168b8eabcbb","Type":"ContainerStarted","Data":"d969915ec6df8b8dbe9c2a65574f09d24938a93579fa72956de4f1f392ca1778"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.619863 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9b9799bc-tvwd9" event={"ID":"76945d28-7017-4b37-ac66-c356e9190230","Type":"ContainerStarted","Data":"17703fd86227714ae35dcfd5c08c07fcf868c83008952e0bb1fde369d0d65166"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.620822 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-545df9b758-drjst" event={"ID":"54742b9b-0346-4cdb-9bf2-75f35c59beaf","Type":"ContainerStarted","Data":"8df35444918c8cffed01719476bcb3e14dc54a478b41ec8bc19fa0de323f6384"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.623891 5035 generic.go:334] "Generic (PLEG): container finished" podID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerID="9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f" exitCode=0 Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.623940 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" event={"ID":"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b","Type":"ContainerDied","Data":"9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.623958 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" event={"ID":"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b","Type":"ContainerStarted","Data":"eb48d2632ef1a6dde91a8991789538bf69baeff8d97595e2b8293d105ac07249"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.631786 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" event={"ID":"d283583a-e964-4817-b474-8c128a4e78b6","Type":"ContainerDied","Data":"31593d12b0ea1d889fcb7cfa4261d6b089b130a719d46841fb0782ca6aaa2c64"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.631836 5035 scope.go:117] "RemoveContainer" containerID="27503ecef9b6e6b735d233e00872ea1d2f8da36bbb55086aa5412dbd44a31c95" Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.631976 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-tv4h4" Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.652196 5035 generic.go:334] "Generic (PLEG): container finished" podID="76ce8ea1-cda3-4f68-9570-c6c02b56283f" containerID="49e6bc1dcd3210e3ae2c9a69fafc0c053a9f33d4a050ddbc3b8504564fdd8373" exitCode=0 Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.652243 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7t2tz" event={"ID":"76ce8ea1-cda3-4f68-9570-c6c02b56283f","Type":"ContainerDied","Data":"49e6bc1dcd3210e3ae2c9a69fafc0c053a9f33d4a050ddbc3b8504564fdd8373"} Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.685702 5035 scope.go:117] "RemoveContainer" containerID="1936d1cec2a4726fe036f2f0c7c249ac4ae357d427f4bfe4d9242a22e0c8f2a0" Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.721232 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tv4h4"] Oct 02 09:45:39 crc kubenswrapper[5035]: I1002 09:45:39.730464 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-tv4h4"] Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.186493 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d283583a-e964-4817-b474-8c128a4e78b6" path="/var/lib/kubelet/pods/d283583a-e964-4817-b474-8c128a4e78b6/volumes" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.695479 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6c775994-6t2fx" event={"ID":"91467c75-7c7d-4b20-baad-b168b8eabcbb","Type":"ContainerStarted","Data":"95bff7fafd8b7f1a7bf03d58dab8a569a8b92450b4beec73a560a8c2852071ba"} Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.697231 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.697277 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.703141 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" event={"ID":"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b","Type":"ContainerStarted","Data":"b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4"} Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.704190 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.728015 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-f6c775994-6t2fx" podStartSLOduration=2.727989852 podStartE2EDuration="2.727989852s" podCreationTimestamp="2025-10-02 09:45:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:40.72549718 +0000 UTC m=+1106.081841205" watchObservedRunningTime="2025-10-02 09:45:40.727989852 +0000 UTC m=+1106.084333887" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.772586 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" podStartSLOduration=3.772564566 podStartE2EDuration="3.772564566s" podCreationTimestamp="2025-10-02 09:45:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:40.753944701 +0000 UTC m=+1106.110288726" watchObservedRunningTime="2025-10-02 09:45:40.772564566 +0000 UTC m=+1106.128908591" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.815025 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7975d57bf8-2hqjm"] Oct 02 09:45:40 crc kubenswrapper[5035]: E1002 09:45:40.815439 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d283583a-e964-4817-b474-8c128a4e78b6" containerName="dnsmasq-dns" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.815457 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d283583a-e964-4817-b474-8c128a4e78b6" containerName="dnsmasq-dns" Oct 02 09:45:40 crc kubenswrapper[5035]: E1002 09:45:40.815469 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d283583a-e964-4817-b474-8c128a4e78b6" containerName="init" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.815476 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="d283583a-e964-4817-b474-8c128a4e78b6" containerName="init" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.815664 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="d283583a-e964-4817-b474-8c128a4e78b6" containerName="dnsmasq-dns" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.816503 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.822764 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.823039 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.843262 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7975d57bf8-2hqjm"] Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.902575 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-config-data-custom\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.902626 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-public-tls-certs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.902666 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-internal-tls-certs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.902691 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwhn6\" (UniqueName: \"kubernetes.io/projected/78fda780-4464-412b-919b-58ee75780eb6-kube-api-access-mwhn6\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.902971 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78fda780-4464-412b-919b-58ee75780eb6-logs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.903030 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-combined-ca-bundle\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:40 crc kubenswrapper[5035]: I1002 09:45:40.903078 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-config-data\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.005152 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-config-data-custom\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.005210 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-public-tls-certs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.005252 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-internal-tls-certs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.005279 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwhn6\" (UniqueName: \"kubernetes.io/projected/78fda780-4464-412b-919b-58ee75780eb6-kube-api-access-mwhn6\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.005349 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78fda780-4464-412b-919b-58ee75780eb6-logs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.005373 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-combined-ca-bundle\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.005395 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-config-data\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.006407 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/78fda780-4464-412b-919b-58ee75780eb6-logs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.029703 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-public-tls-certs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.030906 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-config-data\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.032003 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-internal-tls-certs\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.040170 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwhn6\" (UniqueName: \"kubernetes.io/projected/78fda780-4464-412b-919b-58ee75780eb6-kube-api-access-mwhn6\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.062251 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-combined-ca-bundle\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.063130 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/78fda780-4464-412b-919b-58ee75780eb6-config-data-custom\") pod \"barbican-api-7975d57bf8-2hqjm\" (UID: \"78fda780-4464-412b-919b-58ee75780eb6\") " pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.148478 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.197271 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.315225 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-scripts\") pod \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.315276 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-config-data\") pod \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.315391 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-combined-ca-bundle\") pod \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.315490 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-db-sync-config-data\") pod \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.315571 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpm2v\" (UniqueName: \"kubernetes.io/projected/76ce8ea1-cda3-4f68-9570-c6c02b56283f-kube-api-access-mpm2v\") pod \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.315636 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ce8ea1-cda3-4f68-9570-c6c02b56283f-etc-machine-id\") pod \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\" (UID: \"76ce8ea1-cda3-4f68-9570-c6c02b56283f\") " Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.316297 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76ce8ea1-cda3-4f68-9570-c6c02b56283f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "76ce8ea1-cda3-4f68-9570-c6c02b56283f" (UID: "76ce8ea1-cda3-4f68-9570-c6c02b56283f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.320172 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-scripts" (OuterVolumeSpecName: "scripts") pod "76ce8ea1-cda3-4f68-9570-c6c02b56283f" (UID: "76ce8ea1-cda3-4f68-9570-c6c02b56283f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.323622 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ce8ea1-cda3-4f68-9570-c6c02b56283f-kube-api-access-mpm2v" (OuterVolumeSpecName: "kube-api-access-mpm2v") pod "76ce8ea1-cda3-4f68-9570-c6c02b56283f" (UID: "76ce8ea1-cda3-4f68-9570-c6c02b56283f"). InnerVolumeSpecName "kube-api-access-mpm2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.324983 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "76ce8ea1-cda3-4f68-9570-c6c02b56283f" (UID: "76ce8ea1-cda3-4f68-9570-c6c02b56283f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.366370 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76ce8ea1-cda3-4f68-9570-c6c02b56283f" (UID: "76ce8ea1-cda3-4f68-9570-c6c02b56283f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.417053 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.417081 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.417092 5035 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.417105 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpm2v\" (UniqueName: \"kubernetes.io/projected/76ce8ea1-cda3-4f68-9570-c6c02b56283f-kube-api-access-mpm2v\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.417113 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/76ce8ea1-cda3-4f68-9570-c6c02b56283f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.486365 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-config-data" (OuterVolumeSpecName: "config-data") pod "76ce8ea1-cda3-4f68-9570-c6c02b56283f" (UID: "76ce8ea1-cda3-4f68-9570-c6c02b56283f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.519060 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76ce8ea1-cda3-4f68-9570-c6c02b56283f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.723095 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9b9799bc-tvwd9" event={"ID":"76945d28-7017-4b37-ac66-c356e9190230","Type":"ContainerStarted","Data":"6eec74c7bf521cf66faab1b743f4adbcb3e6a762a76ea604943f743e81a4579a"} Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.732861 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerStarted","Data":"969da5960e52410b4f3c237c7b02dfd11c433dfb2001cc4360fba980463adfe0"} Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.733618 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.739869 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7975d57bf8-2hqjm"] Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.744710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-545df9b758-drjst" event={"ID":"54742b9b-0346-4cdb-9bf2-75f35c59beaf","Type":"ContainerStarted","Data":"c18989ca0afcecbfb629577b03a2239e5a977f2583c404fdfacc38915f1c0683"} Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.766233 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.9896706 podStartE2EDuration="7.766102741s" podCreationTimestamp="2025-10-02 09:45:34 +0000 UTC" firstStartedPulling="2025-10-02 09:45:35.455086521 +0000 UTC m=+1100.811430546" lastFinishedPulling="2025-10-02 09:45:41.231518662 +0000 UTC m=+1106.587862687" observedRunningTime="2025-10-02 09:45:41.758917761 +0000 UTC m=+1107.115261786" watchObservedRunningTime="2025-10-02 09:45:41.766102741 +0000 UTC m=+1107.122446776" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.766874 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7t2tz" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.768711 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7t2tz" event={"ID":"76ce8ea1-cda3-4f68-9570-c6c02b56283f","Type":"ContainerDied","Data":"8ff5d586b5d8e464ecb7fadfe3abe942d5c27229df686a200838c87b458a6587"} Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.768752 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ff5d586b5d8e464ecb7fadfe3abe942d5c27229df686a200838c87b458a6587" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.921879 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:41 crc kubenswrapper[5035]: E1002 09:45:41.922239 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ce8ea1-cda3-4f68-9570-c6c02b56283f" containerName="cinder-db-sync" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.922263 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ce8ea1-cda3-4f68-9570-c6c02b56283f" containerName="cinder-db-sync" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.922675 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ce8ea1-cda3-4f68-9570-c6c02b56283f" containerName="cinder-db-sync" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.934387 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.945324 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-8nccm" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.948880 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.949432 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.949516 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 09:45:41 crc kubenswrapper[5035]: I1002 09:45:41.971715 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.019668 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-ns7jc"] Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.035873 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.036131 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmz88\" (UniqueName: \"kubernetes.io/projected/1af113d9-450d-4644-927a-9c9b122f122a-kube-api-access-gmz88\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.036295 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1af113d9-450d-4644-927a-9c9b122f122a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.036458 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-scripts\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.036584 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.036737 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.087924 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-4f72t"] Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.091148 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.123032 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-4f72t"] Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.137908 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-scripts\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.137957 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.138013 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.138057 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.138135 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmz88\" (UniqueName: \"kubernetes.io/projected/1af113d9-450d-4644-927a-9c9b122f122a-kube-api-access-gmz88\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.138192 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1af113d9-450d-4644-927a-9c9b122f122a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.138343 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1af113d9-450d-4644-927a-9c9b122f122a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.145928 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.146366 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-scripts\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.162977 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.163082 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.177346 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmz88\" (UniqueName: \"kubernetes.io/projected/1af113d9-450d-4644-927a-9c9b122f122a-kube-api-access-gmz88\") pod \"cinder-scheduler-0\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.183972 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.185457 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.185590 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.202375 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.239967 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-config\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.240061 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.240142 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.240215 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgrvz\" (UniqueName: \"kubernetes.io/projected/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-kube-api-access-zgrvz\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.240412 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.240447 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-svc\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.293922 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354056 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354116 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49ltg\" (UniqueName: \"kubernetes.io/projected/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-kube-api-access-49ltg\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354145 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data-custom\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354185 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgrvz\" (UniqueName: \"kubernetes.io/projected/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-kube-api-access-zgrvz\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354256 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-scripts\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354293 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354314 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354334 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-svc\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354364 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-logs\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354402 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-config\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354436 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354455 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.354479 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.355990 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.356161 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.356360 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-svc\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.356690 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-config\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.356912 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.426106 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgrvz\" (UniqueName: \"kubernetes.io/projected/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-kube-api-access-zgrvz\") pod \"dnsmasq-dns-5784cf869f-4f72t\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.455818 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49ltg\" (UniqueName: \"kubernetes.io/projected/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-kube-api-access-49ltg\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.455870 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data-custom\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.455948 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-scripts\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.455996 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.456044 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-logs\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.456125 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.456167 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.460810 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.465204 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-logs\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.465504 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.496463 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.509099 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-scripts\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.509958 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data-custom\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.529845 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49ltg\" (UniqueName: \"kubernetes.io/projected/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-kube-api-access-49ltg\") pod \"cinder-api-0\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.683871 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.747232 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.792102 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5d9b9799bc-tvwd9" event={"ID":"76945d28-7017-4b37-ac66-c356e9190230","Type":"ContainerStarted","Data":"98b0b611972d7213bc4fe157c6b4d77407a6654f9971a08332bb7ce705aa3b53"} Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.812820 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5d9b9799bc-tvwd9" podStartSLOduration=3.4566513 podStartE2EDuration="5.812799541s" podCreationTimestamp="2025-10-02 09:45:37 +0000 UTC" firstStartedPulling="2025-10-02 09:45:38.876435602 +0000 UTC m=+1104.232779617" lastFinishedPulling="2025-10-02 09:45:41.232583833 +0000 UTC m=+1106.588927858" observedRunningTime="2025-10-02 09:45:42.811955176 +0000 UTC m=+1108.168299201" watchObservedRunningTime="2025-10-02 09:45:42.812799541 +0000 UTC m=+1108.169143566" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.815029 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-545df9b758-drjst" event={"ID":"54742b9b-0346-4cdb-9bf2-75f35c59beaf","Type":"ContainerStarted","Data":"bafac6a4abe764acfc26a8de4016dc1cc80411cd1f58b645484ebd36b37e17f4"} Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.822777 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7975d57bf8-2hqjm" event={"ID":"78fda780-4464-412b-919b-58ee75780eb6","Type":"ContainerStarted","Data":"41aeb267f52285e2f473b9d0560e4d8560c0c9a2a7b16abf2cb69ab098998adb"} Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.822818 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7975d57bf8-2hqjm" event={"ID":"78fda780-4464-412b-919b-58ee75780eb6","Type":"ContainerStarted","Data":"8fca455246ee5329c13e11feb2e91ba7edc152affc5019bc40cf268d60d38b6a"} Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.822840 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7975d57bf8-2hqjm" event={"ID":"78fda780-4464-412b-919b-58ee75780eb6","Type":"ContainerStarted","Data":"acc4192a90601f82fa4f492ae83dea852dd9c560e0a7fc169965e69144feebb3"} Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.822855 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.822867 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.844308 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-545df9b758-drjst" podStartSLOduration=3.591827311 podStartE2EDuration="5.844288391s" podCreationTimestamp="2025-10-02 09:45:37 +0000 UTC" firstStartedPulling="2025-10-02 09:45:38.980794283 +0000 UTC m=+1104.337138318" lastFinishedPulling="2025-10-02 09:45:41.233255373 +0000 UTC m=+1106.589599398" observedRunningTime="2025-10-02 09:45:42.838648836 +0000 UTC m=+1108.194992871" watchObservedRunningTime="2025-10-02 09:45:42.844288391 +0000 UTC m=+1108.200632416" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.864406 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7975d57bf8-2hqjm" podStartSLOduration=2.864387099 podStartE2EDuration="2.864387099s" podCreationTimestamp="2025-10-02 09:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:42.864247895 +0000 UTC m=+1108.220591930" watchObservedRunningTime="2025-10-02 09:45:42.864387099 +0000 UTC m=+1108.220731124" Oct 02 09:45:42 crc kubenswrapper[5035]: I1002 09:45:42.969793 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.207215 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-4f72t"] Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.334886 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.841111 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9","Type":"ContainerStarted","Data":"4e8e7e6e9ced818bbd2c8c5881f168fe1956aa3f76fd5add538774d91c7b7882"} Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.843451 5035 generic.go:334] "Generic (PLEG): container finished" podID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerID="50a5c12ef817b3257d4f0e3edf6c3ec5dad480722699ffb96a4fda1c6d53c2a1" exitCode=0 Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.843556 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" event={"ID":"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd","Type":"ContainerDied","Data":"50a5c12ef817b3257d4f0e3edf6c3ec5dad480722699ffb96a4fda1c6d53c2a1"} Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.843584 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" event={"ID":"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd","Type":"ContainerStarted","Data":"a95b8608cdd74b27410b68eefe1a5f18163559361d0cbd547718084a33011303"} Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.850421 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1af113d9-450d-4644-927a-9c9b122f122a","Type":"ContainerStarted","Data":"7959cc01816343aacc49edede813833e82a39e129c9c2bd348f15c7844532d87"} Oct 02 09:45:43 crc kubenswrapper[5035]: I1002 09:45:43.850496 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" podUID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerName="dnsmasq-dns" containerID="cri-o://b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4" gracePeriod=10 Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.411954 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.504809 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-config\") pod \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.504922 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-svc\") pod \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.505023 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-swift-storage-0\") pod \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.505058 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqczm\" (UniqueName: \"kubernetes.io/projected/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-kube-api-access-hqczm\") pod \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.505145 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-sb\") pod \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.505368 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-nb\") pod \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\" (UID: \"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b\") " Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.510706 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-kube-api-access-hqczm" (OuterVolumeSpecName: "kube-api-access-hqczm") pod "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" (UID: "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b"). InnerVolumeSpecName "kube-api-access-hqczm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.610689 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqczm\" (UniqueName: \"kubernetes.io/projected/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-kube-api-access-hqczm\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.623942 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" (UID: "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.631054 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" (UID: "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.639448 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" (UID: "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.666414 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.673752 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-config" (OuterVolumeSpecName: "config") pod "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" (UID: "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.673913 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" (UID: "6cccbab7-37f6-4826-b3d3-c110f7ba0f4b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.713202 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.713233 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.713245 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.713253 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.713262 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.860202 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1af113d9-450d-4644-927a-9c9b122f122a","Type":"ContainerStarted","Data":"e948f901034ac36e70079a8d9a37da4905b906e22946fca778dfb603ddc50e86"} Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.866370 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9","Type":"ContainerStarted","Data":"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15"} Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.869118 5035 generic.go:334] "Generic (PLEG): container finished" podID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerID="b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4" exitCode=0 Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.869162 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" event={"ID":"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b","Type":"ContainerDied","Data":"b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4"} Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.869179 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" event={"ID":"6cccbab7-37f6-4826-b3d3-c110f7ba0f4b","Type":"ContainerDied","Data":"eb48d2632ef1a6dde91a8991789538bf69baeff8d97595e2b8293d105ac07249"} Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.869195 5035 scope.go:117] "RemoveContainer" containerID="b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.869308 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-ns7jc" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.885741 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" event={"ID":"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd","Type":"ContainerStarted","Data":"a10d5e846c8ea3f7112b3a156244cc8d64d57eeb3b7d582c28802531b8587bbe"} Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.886848 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.927768 5035 scope.go:117] "RemoveContainer" containerID="9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.930934 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" podStartSLOduration=3.930916463 podStartE2EDuration="3.930916463s" podCreationTimestamp="2025-10-02 09:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:44.90757014 +0000 UTC m=+1110.263914165" watchObservedRunningTime="2025-10-02 09:45:44.930916463 +0000 UTC m=+1110.287260488" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.948618 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-ns7jc"] Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.965512 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-ns7jc"] Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.993496 5035 scope.go:117] "RemoveContainer" containerID="b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4" Oct 02 09:45:44 crc kubenswrapper[5035]: E1002 09:45:44.997691 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4\": container with ID starting with b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4 not found: ID does not exist" containerID="b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.997733 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4"} err="failed to get container status \"b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4\": rpc error: code = NotFound desc = could not find container \"b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4\": container with ID starting with b360c63f5c830f0dda164046919c71ef332465a0a58997ca412c727ed16941c4 not found: ID does not exist" Oct 02 09:45:44 crc kubenswrapper[5035]: I1002 09:45:44.997758 5035 scope.go:117] "RemoveContainer" containerID="9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f" Oct 02 09:45:45 crc kubenswrapper[5035]: E1002 09:45:45.005705 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f\": container with ID starting with 9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f not found: ID does not exist" containerID="9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f" Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.006033 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f"} err="failed to get container status \"9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f\": rpc error: code = NotFound desc = could not find container \"9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f\": container with ID starting with 9fb8e43a3630bedc297e8584c30dbd28a618eaced58b6a4bf932fb71ce2d073f not found: ID does not exist" Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.896615 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1af113d9-450d-4644-927a-9c9b122f122a","Type":"ContainerStarted","Data":"83ee2d74a47a74610498a0751f0b680ead41b43110840535555680cee075f32b"} Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.898639 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9","Type":"ContainerStarted","Data":"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40"} Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.898779 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api-log" containerID="cri-o://68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15" gracePeriod=30 Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.898835 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api" containerID="cri-o://38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40" gracePeriod=30 Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.898846 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.944320 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.082774032 podStartE2EDuration="4.944295129s" podCreationTimestamp="2025-10-02 09:45:41 +0000 UTC" firstStartedPulling="2025-10-02 09:45:43.02931204 +0000 UTC m=+1108.385656055" lastFinishedPulling="2025-10-02 09:45:43.890833127 +0000 UTC m=+1109.247177152" observedRunningTime="2025-10-02 09:45:45.922890513 +0000 UTC m=+1111.279234548" watchObservedRunningTime="2025-10-02 09:45:45.944295129 +0000 UTC m=+1111.300639154" Oct 02 09:45:45 crc kubenswrapper[5035]: I1002 09:45:45.947811 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.947791471 podStartE2EDuration="3.947791471s" podCreationTimestamp="2025-10-02 09:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:45.940807127 +0000 UTC m=+1111.297151162" watchObservedRunningTime="2025-10-02 09:45:45.947791471 +0000 UTC m=+1111.304135496" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.178435 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" path="/var/lib/kubelet/pods/6cccbab7-37f6-4826-b3d3-c110f7ba0f4b/volumes" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.474049 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.553056 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49ltg\" (UniqueName: \"kubernetes.io/projected/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-kube-api-access-49ltg\") pod \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.553101 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-logs\") pod \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.553140 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-combined-ca-bundle\") pod \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.553200 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-etc-machine-id\") pod \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.553282 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data-custom\") pod \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.553295 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data\") pod \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.553311 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-scripts\") pod \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\" (UID: \"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9\") " Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.555051 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-logs" (OuterVolumeSpecName: "logs") pod "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" (UID: "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.555119 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" (UID: "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.561943 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" (UID: "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.562743 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-kube-api-access-49ltg" (OuterVolumeSpecName: "kube-api-access-49ltg") pod "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" (UID: "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9"). InnerVolumeSpecName "kube-api-access-49ltg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.562880 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-scripts" (OuterVolumeSpecName: "scripts") pod "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" (UID: "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.586718 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" (UID: "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.618895 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data" (OuterVolumeSpecName: "config-data") pod "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" (UID: "0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.654947 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49ltg\" (UniqueName: \"kubernetes.io/projected/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-kube-api-access-49ltg\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.654983 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.654993 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.655001 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.655011 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.655019 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.655027 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.913134 5035 generic.go:334] "Generic (PLEG): container finished" podID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerID="38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40" exitCode=0 Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.913166 5035 generic.go:334] "Generic (PLEG): container finished" podID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerID="68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15" exitCode=143 Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.913196 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.913200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9","Type":"ContainerDied","Data":"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40"} Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.913256 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9","Type":"ContainerDied","Data":"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15"} Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.913269 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9","Type":"ContainerDied","Data":"4e8e7e6e9ced818bbd2c8c5881f168fe1956aa3f76fd5add538774d91c7b7882"} Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.913291 5035 scope.go:117] "RemoveContainer" containerID="38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.915884 5035 generic.go:334] "Generic (PLEG): container finished" podID="afad3fdc-5375-4bba-9832-f94381ba82aa" containerID="9ed1c71145077c8e62f1a887d9e9e42bc8652299ccafecb1821b2fdf53bc90d5" exitCode=0 Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.915966 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-klp2h" event={"ID":"afad3fdc-5375-4bba-9832-f94381ba82aa","Type":"ContainerDied","Data":"9ed1c71145077c8e62f1a887d9e9e42bc8652299ccafecb1821b2fdf53bc90d5"} Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.955839 5035 scope.go:117] "RemoveContainer" containerID="68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.963675 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.968731 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.989207 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:46 crc kubenswrapper[5035]: E1002 09:45:46.989587 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api-log" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.989602 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api-log" Oct 02 09:45:46 crc kubenswrapper[5035]: E1002 09:45:46.989623 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerName="init" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.989629 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerName="init" Oct 02 09:45:46 crc kubenswrapper[5035]: E1002 09:45:46.989640 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerName="dnsmasq-dns" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.989645 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerName="dnsmasq-dns" Oct 02 09:45:46 crc kubenswrapper[5035]: E1002 09:45:46.989664 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.989669 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.990898 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cccbab7-37f6-4826-b3d3-c110f7ba0f4b" containerName="dnsmasq-dns" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.990916 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api-log" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.990935 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" containerName="cinder-api" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.991924 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.992908 5035 scope.go:117] "RemoveContainer" containerID="38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.995998 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.996392 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.996675 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 09:45:46 crc kubenswrapper[5035]: E1002 09:45:46.999241 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40\": container with ID starting with 38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40 not found: ID does not exist" containerID="38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.999290 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40"} err="failed to get container status \"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40\": rpc error: code = NotFound desc = could not find container \"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40\": container with ID starting with 38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40 not found: ID does not exist" Oct 02 09:45:46 crc kubenswrapper[5035]: I1002 09:45:46.999321 5035 scope.go:117] "RemoveContainer" containerID="68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15" Oct 02 09:45:47 crc kubenswrapper[5035]: E1002 09:45:47.002577 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15\": container with ID starting with 68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15 not found: ID does not exist" containerID="68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.002880 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15"} err="failed to get container status \"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15\": rpc error: code = NotFound desc = could not find container \"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15\": container with ID starting with 68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15 not found: ID does not exist" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.002901 5035 scope.go:117] "RemoveContainer" containerID="38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.003639 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40"} err="failed to get container status \"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40\": rpc error: code = NotFound desc = could not find container \"38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40\": container with ID starting with 38199b5003db9d9eb2abf9cd5805539a0e9e8d5356f3726f2c3b275e60d6ab40 not found: ID does not exist" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.003662 5035 scope.go:117] "RemoveContainer" containerID="68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.004291 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15"} err="failed to get container status \"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15\": rpc error: code = NotFound desc = could not find container \"68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15\": container with ID starting with 68c93bcc1041930f259c3c1b80c37df39157e257bea387935e663a10de49ab15 not found: ID does not exist" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.006388 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060415 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060483 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060510 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060716 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060754 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-config-data\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060881 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060911 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-scripts\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.060971 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n29vr\" (UniqueName: \"kubernetes.io/projected/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-kube-api-access-n29vr\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.061009 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-logs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163064 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163173 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163192 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163257 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163305 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163328 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-config-data\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163409 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163433 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-scripts\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163471 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n29vr\" (UniqueName: \"kubernetes.io/projected/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-kube-api-access-n29vr\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.163510 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-logs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.164014 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-logs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.167557 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.168169 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.168626 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-config-data\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.170946 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-scripts\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.171090 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.178983 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.185772 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n29vr\" (UniqueName: \"kubernetes.io/projected/8ccd0633-b703-46d1-b74c-9129d3b5b9ac-kube-api-access-n29vr\") pod \"cinder-api-0\" (UID: \"8ccd0633-b703-46d1-b74c-9129d3b5b9ac\") " pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.295143 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.313064 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.779050 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 09:45:47 crc kubenswrapper[5035]: W1002 09:45:47.785257 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ccd0633_b703_46d1_b74c_9129d3b5b9ac.slice/crio-f3e7d53cecf9112397e688dafe2887c321e187861e9e5b7c596dae07a87de004 WatchSource:0}: Error finding container f3e7d53cecf9112397e688dafe2887c321e187861e9e5b7c596dae07a87de004: Status 404 returned error can't find the container with id f3e7d53cecf9112397e688dafe2887c321e187861e9e5b7c596dae07a87de004 Oct 02 09:45:47 crc kubenswrapper[5035]: I1002 09:45:47.925260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ccd0633-b703-46d1-b74c-9129d3b5b9ac","Type":"ContainerStarted","Data":"f3e7d53cecf9112397e688dafe2887c321e187861e9e5b7c596dae07a87de004"} Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.175050 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9" path="/var/lib/kubelet/pods/0f1e5893-496c-4a03-bf9d-b6d09d1fa6f9/volumes" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.279028 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.382776 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-combined-ca-bundle\") pod \"afad3fdc-5375-4bba-9832-f94381ba82aa\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.382870 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data\") pod \"afad3fdc-5375-4bba-9832-f94381ba82aa\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.382894 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data-merged\") pod \"afad3fdc-5375-4bba-9832-f94381ba82aa\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.382937 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-scripts\") pod \"afad3fdc-5375-4bba-9832-f94381ba82aa\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.383364 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "afad3fdc-5375-4bba-9832-f94381ba82aa" (UID: "afad3fdc-5375-4bba-9832-f94381ba82aa"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.383469 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spqqn\" (UniqueName: \"kubernetes.io/projected/afad3fdc-5375-4bba-9832-f94381ba82aa-kube-api-access-spqqn\") pod \"afad3fdc-5375-4bba-9832-f94381ba82aa\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.383504 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/afad3fdc-5375-4bba-9832-f94381ba82aa-etc-podinfo\") pod \"afad3fdc-5375-4bba-9832-f94381ba82aa\" (UID: \"afad3fdc-5375-4bba-9832-f94381ba82aa\") " Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.384302 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.390248 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afad3fdc-5375-4bba-9832-f94381ba82aa-kube-api-access-spqqn" (OuterVolumeSpecName: "kube-api-access-spqqn") pod "afad3fdc-5375-4bba-9832-f94381ba82aa" (UID: "afad3fdc-5375-4bba-9832-f94381ba82aa"). InnerVolumeSpecName "kube-api-access-spqqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.407911 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/afad3fdc-5375-4bba-9832-f94381ba82aa-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "afad3fdc-5375-4bba-9832-f94381ba82aa" (UID: "afad3fdc-5375-4bba-9832-f94381ba82aa"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.409621 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-scripts" (OuterVolumeSpecName: "scripts") pod "afad3fdc-5375-4bba-9832-f94381ba82aa" (UID: "afad3fdc-5375-4bba-9832-f94381ba82aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.428790 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data" (OuterVolumeSpecName: "config-data") pod "afad3fdc-5375-4bba-9832-f94381ba82aa" (UID: "afad3fdc-5375-4bba-9832-f94381ba82aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.486123 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spqqn\" (UniqueName: \"kubernetes.io/projected/afad3fdc-5375-4bba-9832-f94381ba82aa-kube-api-access-spqqn\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.486408 5035 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/afad3fdc-5375-4bba-9832-f94381ba82aa-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.486418 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.486426 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.502825 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afad3fdc-5375-4bba-9832-f94381ba82aa" (UID: "afad3fdc-5375-4bba-9832-f94381ba82aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.588401 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afad3fdc-5375-4bba-9832-f94381ba82aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.961483 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-klp2h" event={"ID":"afad3fdc-5375-4bba-9832-f94381ba82aa","Type":"ContainerDied","Data":"c27323c7e57c6556dc856f0f30e811673f692f7309ba9c7db35a02c8e13399ef"} Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.961551 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c27323c7e57c6556dc856f0f30e811673f692f7309ba9c7db35a02c8e13399ef" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.961688 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-klp2h" Oct 02 09:45:48 crc kubenswrapper[5035]: I1002 09:45:48.973265 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ccd0633-b703-46d1-b74c-9129d3b5b9ac","Type":"ContainerStarted","Data":"47ed8cdd20e8668ea8b9b236e376b8f2ad09ebfb35c669b5cf38d9e8a0494047"} Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.291591 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-create-l42l7"] Oct 02 09:45:49 crc kubenswrapper[5035]: E1002 09:45:49.292260 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afad3fdc-5375-4bba-9832-f94381ba82aa" containerName="ironic-db-sync" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.292277 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="afad3fdc-5375-4bba-9832-f94381ba82aa" containerName="ironic-db-sync" Oct 02 09:45:49 crc kubenswrapper[5035]: E1002 09:45:49.292297 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afad3fdc-5375-4bba-9832-f94381ba82aa" containerName="init" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.292304 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="afad3fdc-5375-4bba-9832-f94381ba82aa" containerName="init" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.292467 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="afad3fdc-5375-4bba-9832-f94381ba82aa" containerName="ironic-db-sync" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.293594 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-l42l7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.314788 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-l42l7"] Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.358866 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-neutron-agent-679cc45cbb-lgkg7"] Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.360153 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.365624 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-8t5qg" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.365881 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-neutron-agent-config-data" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.393801 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-679cc45cbb-lgkg7"] Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.401915 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z9zb\" (UniqueName: \"kubernetes.io/projected/9fec464c-ae40-4d5f-8303-ed2c93babb81-kube-api-access-8z9zb\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.402199 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fec464c-ae40-4d5f-8303-ed2c93babb81-config\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.402307 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fec464c-ae40-4d5f-8303-ed2c93babb81-combined-ca-bundle\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.402440 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffs7k\" (UniqueName: \"kubernetes.io/projected/7e575f2f-c857-4a4f-a771-e591df208a9c-kube-api-access-ffs7k\") pod \"ironic-inspector-db-create-l42l7\" (UID: \"7e575f2f-c857-4a4f-a771-e591df208a9c\") " pod="openstack/ironic-inspector-db-create-l42l7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.503747 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fec464c-ae40-4d5f-8303-ed2c93babb81-combined-ca-bundle\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.503869 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffs7k\" (UniqueName: \"kubernetes.io/projected/7e575f2f-c857-4a4f-a771-e591df208a9c-kube-api-access-ffs7k\") pod \"ironic-inspector-db-create-l42l7\" (UID: \"7e575f2f-c857-4a4f-a771-e591df208a9c\") " pod="openstack/ironic-inspector-db-create-l42l7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.503980 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z9zb\" (UniqueName: \"kubernetes.io/projected/9fec464c-ae40-4d5f-8303-ed2c93babb81-kube-api-access-8z9zb\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.504035 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fec464c-ae40-4d5f-8303-ed2c93babb81-config\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.524758 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fec464c-ae40-4d5f-8303-ed2c93babb81-combined-ca-bundle\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.526019 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9fec464c-ae40-4d5f-8303-ed2c93babb81-config\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.528664 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffs7k\" (UniqueName: \"kubernetes.io/projected/7e575f2f-c857-4a4f-a771-e591df208a9c-kube-api-access-ffs7k\") pod \"ironic-inspector-db-create-l42l7\" (UID: \"7e575f2f-c857-4a4f-a771-e591df208a9c\") " pod="openstack/ironic-inspector-db-create-l42l7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.543760 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z9zb\" (UniqueName: \"kubernetes.io/projected/9fec464c-ae40-4d5f-8303-ed2c93babb81-kube-api-access-8z9zb\") pod \"ironic-neutron-agent-679cc45cbb-lgkg7\" (UID: \"9fec464c-ae40-4d5f-8303-ed2c93babb81\") " pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.618693 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-l42l7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.652636 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-7c84bc89d8-vz98z"] Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.671010 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.677934 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-config-data" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.678137 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.679713 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-scripts" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.682556 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-7c84bc89d8-vz98z"] Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.692944 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.708724 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-logs\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.708787 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-custom\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.708859 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/bdfd7898-4f4c-4270-9be3-0439621a01a9-etc-podinfo\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.708911 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cqbc\" (UniqueName: \"kubernetes.io/projected/bdfd7898-4f4c-4270-9be3-0439621a01a9-kube-api-access-7cqbc\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.709043 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-scripts\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.709103 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.709145 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-merged\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.709214 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-combined-ca-bundle\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811343 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-scripts\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811392 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811414 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-merged\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811443 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-combined-ca-bundle\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811550 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-logs\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811573 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-custom\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811594 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/bdfd7898-4f4c-4270-9be3-0439621a01a9-etc-podinfo\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.811623 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cqbc\" (UniqueName: \"kubernetes.io/projected/bdfd7898-4f4c-4270-9be3-0439621a01a9-kube-api-access-7cqbc\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.812096 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-logs\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.812347 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-merged\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.818457 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-scripts\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.819059 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-combined-ca-bundle\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.821588 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/bdfd7898-4f4c-4270-9be3-0439621a01a9-etc-podinfo\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.838019 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.838096 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-custom\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.838694 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cqbc\" (UniqueName: \"kubernetes.io/projected/bdfd7898-4f4c-4270-9be3-0439621a01a9-kube-api-access-7cqbc\") pod \"ironic-7c84bc89d8-vz98z\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:49 crc kubenswrapper[5035]: I1002 09:45:49.998790 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ccd0633-b703-46d1-b74c-9129d3b5b9ac","Type":"ContainerStarted","Data":"036c898f95c859c05cb9ec57936ed5ef1e55531f5f96a74ccb0dc2520b37a612"} Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.000462 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.027621 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.027593283 podStartE2EDuration="4.027593283s" podCreationTimestamp="2025-10-02 09:45:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:45:50.022647988 +0000 UTC m=+1115.378992053" watchObservedRunningTime="2025-10-02 09:45:50.027593283 +0000 UTC m=+1115.383937318" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.096648 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.128782 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-l42l7"] Oct 02 09:45:50 crc kubenswrapper[5035]: W1002 09:45:50.134598 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e575f2f_c857_4a4f_a771_e591df208a9c.slice/crio-6060f5331be4af7e8341fc6ea21aea38d29b41b3af34fb5663f0037cc6d3d709 WatchSource:0}: Error finding container 6060f5331be4af7e8341fc6ea21aea38d29b41b3af34fb5663f0037cc6d3d709: Status 404 returned error can't find the container with id 6060f5331be4af7e8341fc6ea21aea38d29b41b3af34fb5663f0037cc6d3d709 Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.286470 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-679cc45cbb-lgkg7"] Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.410599 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-conductor-0"] Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.425756 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.429980 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-scripts" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.430233 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-config-data" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.446690 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544178 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjj49\" (UniqueName: \"kubernetes.io/projected/0307b78d-517d-4442-a0d7-1efa3f46098f-kube-api-access-kjj49\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544257 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544297 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544348 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544412 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544449 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0307b78d-517d-4442-a0d7-1efa3f46098f-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544520 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.544591 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-scripts\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.571823 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646010 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-scripts\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646075 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjj49\" (UniqueName: \"kubernetes.io/projected/0307b78d-517d-4442-a0d7-1efa3f46098f-kube-api-access-kjj49\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646113 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646136 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646165 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646235 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646282 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0307b78d-517d-4442-a0d7-1efa3f46098f-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646363 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.646846 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.651294 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.666248 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0307b78d-517d-4442-a0d7-1efa3f46098f-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.684511 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.684991 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-scripts\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.691068 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.691502 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0307b78d-517d-4442-a0d7-1efa3f46098f-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.730110 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjj49\" (UniqueName: \"kubernetes.io/projected/0307b78d-517d-4442-a0d7-1efa3f46098f-kube-api-access-kjj49\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.806816 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ironic-conductor-0\" (UID: \"0307b78d-517d-4442-a0d7-1efa3f46098f\") " pod="openstack/ironic-conductor-0" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.882913 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:50 crc kubenswrapper[5035]: I1002 09:45:50.925513 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-7c84bc89d8-vz98z"] Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.016962 5035 generic.go:334] "Generic (PLEG): container finished" podID="7e575f2f-c857-4a4f-a771-e591df208a9c" containerID="a3c54c4b74f2501aebce1fdc725b64e3df3f6281017bdbc507ead15479f2b2b5" exitCode=0 Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.017019 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-l42l7" event={"ID":"7e575f2f-c857-4a4f-a771-e591df208a9c","Type":"ContainerDied","Data":"a3c54c4b74f2501aebce1fdc725b64e3df3f6281017bdbc507ead15479f2b2b5"} Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.017045 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-l42l7" event={"ID":"7e575f2f-c857-4a4f-a771-e591df208a9c","Type":"ContainerStarted","Data":"6060f5331be4af7e8341fc6ea21aea38d29b41b3af34fb5663f0037cc6d3d709"} Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.026734 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" event={"ID":"9fec464c-ae40-4d5f-8303-ed2c93babb81","Type":"ContainerStarted","Data":"93d7a18fb41952f6425a654c7b371c889449056ea47941e05daa7317fc5f7f7e"} Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.030204 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerStarted","Data":"3862e6c597c640e753e6595104433d48f116779719562b0fac4098942b9cb6e8"} Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.056010 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.740817 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Oct 02 09:45:51 crc kubenswrapper[5035]: I1002 09:45:51.818868 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.099824 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerStarted","Data":"366b99d5f8cffcee6158acbff2d2d44a83ed372447ab785a9147484e42dd6035"} Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.299026 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57bc66b65b-7wmtt" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.519206 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-7b996fb4b6-gvrhc"] Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.521210 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.525054 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-internal-svc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.525340 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-public-svc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.538067 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-7b996fb4b6-gvrhc"] Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634490 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-scripts\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634608 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jxzn\" (UniqueName: \"kubernetes.io/projected/c596018f-cc4e-4dc4-8c37-def6f2b19f94-kube-api-access-9jxzn\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634636 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c596018f-cc4e-4dc4-8c37-def6f2b19f94-etc-podinfo\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634677 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-combined-ca-bundle\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634706 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-public-tls-certs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634723 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634746 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-internal-tls-certs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634765 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c596018f-cc4e-4dc4-8c37-def6f2b19f94-logs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.634796 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data-merged\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.635024 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data-custom\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.685674 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.726882 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-l42l7" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-combined-ca-bundle\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740246 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-public-tls-certs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740289 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-internal-tls-certs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740305 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c596018f-cc4e-4dc4-8c37-def6f2b19f94-logs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740381 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data-merged\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740418 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data-custom\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-scripts\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740516 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jxzn\" (UniqueName: \"kubernetes.io/projected/c596018f-cc4e-4dc4-8c37-def6f2b19f94-kube-api-access-9jxzn\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.740553 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c596018f-cc4e-4dc4-8c37-def6f2b19f94-etc-podinfo\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.754961 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data-merged\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.760466 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c596018f-cc4e-4dc4-8c37-def6f2b19f94-etc-podinfo\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.761722 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-combined-ca-bundle\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.761768 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c596018f-cc4e-4dc4-8c37-def6f2b19f94-logs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.773274 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-internal-tls-certs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.773715 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-scripts\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.774225 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-public-tls-certs\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.775502 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.803274 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-286pl"] Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.803829 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" podUID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerName="dnsmasq-dns" containerID="cri-o://bb0957b6c8828276a9c0dbdb0b34512261a2a5c3c1c350435b3a4024026c18f2" gracePeriod=10 Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.839415 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jxzn\" (UniqueName: \"kubernetes.io/projected/c596018f-cc4e-4dc4-8c37-def6f2b19f94-kube-api-access-9jxzn\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.858401 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffs7k\" (UniqueName: \"kubernetes.io/projected/7e575f2f-c857-4a4f-a771-e591df208a9c-kube-api-access-ffs7k\") pod \"7e575f2f-c857-4a4f-a771-e591df208a9c\" (UID: \"7e575f2f-c857-4a4f-a771-e591df208a9c\") " Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.867411 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e575f2f-c857-4a4f-a771-e591df208a9c-kube-api-access-ffs7k" (OuterVolumeSpecName: "kube-api-access-ffs7k") pod "7e575f2f-c857-4a4f-a771-e591df208a9c" (UID: "7e575f2f-c857-4a4f-a771-e591df208a9c"). InnerVolumeSpecName "kube-api-access-ffs7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.871506 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c596018f-cc4e-4dc4-8c37-def6f2b19f94-config-data-custom\") pod \"ironic-7b996fb4b6-gvrhc\" (UID: \"c596018f-cc4e-4dc4-8c37-def6f2b19f94\") " pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.907950 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-59898c4b58-pz6vt" Oct 02 09:45:52 crc kubenswrapper[5035]: I1002 09:45:52.961981 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffs7k\" (UniqueName: \"kubernetes.io/projected/7e575f2f-c857-4a4f-a771-e591df208a9c-kube-api-access-ffs7k\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.009073 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.015078 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.078220 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.142241 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" event={"ID":"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45","Type":"ContainerDied","Data":"bb0957b6c8828276a9c0dbdb0b34512261a2a5c3c1c350435b3a4024026c18f2"} Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.143519 5035 generic.go:334] "Generic (PLEG): container finished" podID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerID="bb0957b6c8828276a9c0dbdb0b34512261a2a5c3c1c350435b3a4024026c18f2" exitCode=0 Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.191061 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerStarted","Data":"d41a70bb4640fbbb4aec6b964a17d10bc425a65d9b5f3143a7202d9d6e585844"} Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.214037 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="cinder-scheduler" containerID="cri-o://e948f901034ac36e70079a8d9a37da4905b906e22946fca778dfb603ddc50e86" gracePeriod=30 Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.214361 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-l42l7" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.215258 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-l42l7" event={"ID":"7e575f2f-c857-4a4f-a771-e591df208a9c","Type":"ContainerDied","Data":"6060f5331be4af7e8341fc6ea21aea38d29b41b3af34fb5663f0037cc6d3d709"} Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.215287 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6060f5331be4af7e8341fc6ea21aea38d29b41b3af34fb5663f0037cc6d3d709" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.215322 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="probe" containerID="cri-o://83ee2d74a47a74610498a0751f0b680ead41b43110840535555680cee075f32b" gracePeriod=30 Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.414612 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.630581 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.674554 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-7b996fb4b6-gvrhc"] Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.677863 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-svc\") pod \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.677926 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-config\") pod \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.677985 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5q6g\" (UniqueName: \"kubernetes.io/projected/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-kube-api-access-z5q6g\") pod \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.678028 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-nb\") pod \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.678065 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-swift-storage-0\") pod \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.678105 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-sb\") pod \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\" (UID: \"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45\") " Oct 02 09:45:53 crc kubenswrapper[5035]: W1002 09:45:53.693697 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc596018f_cc4e_4dc4_8c37_def6f2b19f94.slice/crio-e6e81febabc0fa5e0d4625d835910978198d3c3ce436c0c0ae64b437787d34ab WatchSource:0}: Error finding container e6e81febabc0fa5e0d4625d835910978198d3c3ce436c0c0ae64b437787d34ab: Status 404 returned error can't find the container with id e6e81febabc0fa5e0d4625d835910978198d3c3ce436c0c0ae64b437787d34ab Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.708043 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-kube-api-access-z5q6g" (OuterVolumeSpecName: "kube-api-access-z5q6g") pod "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" (UID: "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45"). InnerVolumeSpecName "kube-api-access-z5q6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.747282 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 09:45:53 crc kubenswrapper[5035]: E1002 09:45:53.748034 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e575f2f-c857-4a4f-a771-e591df208a9c" containerName="mariadb-database-create" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.748057 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e575f2f-c857-4a4f-a771-e591df208a9c" containerName="mariadb-database-create" Oct 02 09:45:53 crc kubenswrapper[5035]: E1002 09:45:53.748072 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerName="dnsmasq-dns" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.748082 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerName="dnsmasq-dns" Oct 02 09:45:53 crc kubenswrapper[5035]: E1002 09:45:53.748110 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerName="init" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.748118 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerName="init" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.748366 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e575f2f-c857-4a4f-a771-e591df208a9c" containerName="mariadb-database-create" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.748414 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" containerName="dnsmasq-dns" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.749176 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.756390 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.756678 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-j4zbk" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.756846 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.765953 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.777229 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-config" (OuterVolumeSpecName: "config") pod "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" (UID: "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.786452 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b71f3f-e46f-4025-9935-e000c44652b5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.786591 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/68b71f3f-e46f-4025-9935-e000c44652b5-openstack-config\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.786700 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68b71f3f-e46f-4025-9935-e000c44652b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.786913 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhns7\" (UniqueName: \"kubernetes.io/projected/68b71f3f-e46f-4025-9935-e000c44652b5-kube-api-access-rhns7\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.787096 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5q6g\" (UniqueName: \"kubernetes.io/projected/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-kube-api-access-z5q6g\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.787122 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.795074 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" (UID: "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.812942 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" (UID: "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.814299 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" (UID: "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.828001 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" (UID: "288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889026 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b71f3f-e46f-4025-9935-e000c44652b5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889179 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/68b71f3f-e46f-4025-9935-e000c44652b5-openstack-config\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889318 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68b71f3f-e46f-4025-9935-e000c44652b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889389 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhns7\" (UniqueName: \"kubernetes.io/projected/68b71f3f-e46f-4025-9935-e000c44652b5-kube-api-access-rhns7\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889449 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889464 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889476 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.889488 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.894658 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68b71f3f-e46f-4025-9935-e000c44652b5-combined-ca-bundle\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.895673 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/68b71f3f-e46f-4025-9935-e000c44652b5-openstack-config\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.901034 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/68b71f3f-e46f-4025-9935-e000c44652b5-openstack-config-secret\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:53 crc kubenswrapper[5035]: I1002 09:45:53.909134 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhns7\" (UniqueName: \"kubernetes.io/projected/68b71f3f-e46f-4025-9935-e000c44652b5-kube-api-access-rhns7\") pod \"openstackclient\" (UID: \"68b71f3f-e46f-4025-9935-e000c44652b5\") " pod="openstack/openstackclient" Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.148161 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.228435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7b996fb4b6-gvrhc" event={"ID":"c596018f-cc4e-4dc4-8c37-def6f2b19f94","Type":"ContainerStarted","Data":"e6e81febabc0fa5e0d4625d835910978198d3c3ce436c0c0ae64b437787d34ab"} Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.233680 5035 generic.go:334] "Generic (PLEG): container finished" podID="0307b78d-517d-4442-a0d7-1efa3f46098f" containerID="d41a70bb4640fbbb4aec6b964a17d10bc425a65d9b5f3143a7202d9d6e585844" exitCode=0 Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.233768 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerDied","Data":"d41a70bb4640fbbb4aec6b964a17d10bc425a65d9b5f3143a7202d9d6e585844"} Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.236972 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" event={"ID":"288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45","Type":"ContainerDied","Data":"71d46c208f70aaeb8f8a9d17c92821c2c3991c4defacdc07151ab15238017960"} Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.237018 5035 scope.go:117] "RemoveContainer" containerID="bb0957b6c8828276a9c0dbdb0b34512261a2a5c3c1c350435b3a4024026c18f2" Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.237123 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-286pl" Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.251256 5035 generic.go:334] "Generic (PLEG): container finished" podID="1af113d9-450d-4644-927a-9c9b122f122a" containerID="83ee2d74a47a74610498a0751f0b680ead41b43110840535555680cee075f32b" exitCode=0 Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.251302 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1af113d9-450d-4644-927a-9c9b122f122a","Type":"ContainerDied","Data":"83ee2d74a47a74610498a0751f0b680ead41b43110840535555680cee075f32b"} Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.312839 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-286pl"] Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.318843 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-286pl"] Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.473118 5035 scope.go:117] "RemoveContainer" containerID="0d0a7a63acd78ac35565d708cfecfec5228b463d489da782ac260ec7c5e708a8" Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.596514 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7975d57bf8-2hqjm" Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.661463 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f6c775994-6t2fx"] Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.669568 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f6c775994-6t2fx" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api-log" containerID="cri-o://d93282a16895c929b93ba1290144e9be905106614639e0cfc8df3cb492d7061e" gracePeriod=30 Oct 02 09:45:54 crc kubenswrapper[5035]: I1002 09:45:54.670066 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-f6c775994-6t2fx" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api" containerID="cri-o://95bff7fafd8b7f1a7bf03d58dab8a569a8b92450b4beec73a560a8c2852071ba" gracePeriod=30 Oct 02 09:45:55 crc kubenswrapper[5035]: I1002 09:45:55.265522 5035 generic.go:334] "Generic (PLEG): container finished" podID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerID="d93282a16895c929b93ba1290144e9be905106614639e0cfc8df3cb492d7061e" exitCode=143 Oct 02 09:45:55 crc kubenswrapper[5035]: I1002 09:45:55.265735 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6c775994-6t2fx" event={"ID":"91467c75-7c7d-4b20-baad-b168b8eabcbb","Type":"ContainerDied","Data":"d93282a16895c929b93ba1290144e9be905106614639e0cfc8df3cb492d7061e"} Oct 02 09:45:55 crc kubenswrapper[5035]: I1002 09:45:55.538651 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:45:55 crc kubenswrapper[5035]: I1002 09:45:55.539188 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.013836 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 09:45:56 crc kubenswrapper[5035]: W1002 09:45:56.022444 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68b71f3f_e46f_4025_9935_e000c44652b5.slice/crio-902aee7c306d5b4d2a551f34367ea97b824965977df733b2462d45e6ea7ba43e WatchSource:0}: Error finding container 902aee7c306d5b4d2a551f34367ea97b824965977df733b2462d45e6ea7ba43e: Status 404 returned error can't find the container with id 902aee7c306d5b4d2a551f34367ea97b824965977df733b2462d45e6ea7ba43e Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.178045 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45" path="/var/lib/kubelet/pods/288b9d42-8b9c-4fa7-a1d6-ebfdce76bb45/volumes" Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.277471 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerStarted","Data":"c00b69ba896d36be0e593525a5c9613e9fdb290d350efd1af94d59d38489e81f"} Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.283143 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7b996fb4b6-gvrhc" event={"ID":"c596018f-cc4e-4dc4-8c37-def6f2b19f94","Type":"ContainerStarted","Data":"fd916db38b6c434f51c2e8bde4ef07b05bd231ff988eaab3e3052ea48e9b33e3"} Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.287418 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"68b71f3f-e46f-4025-9935-e000c44652b5","Type":"ContainerStarted","Data":"902aee7c306d5b4d2a551f34367ea97b824965977df733b2462d45e6ea7ba43e"} Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.289357 5035 generic.go:334] "Generic (PLEG): container finished" podID="1af113d9-450d-4644-927a-9c9b122f122a" containerID="e948f901034ac36e70079a8d9a37da4905b906e22946fca778dfb603ddc50e86" exitCode=0 Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.289406 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1af113d9-450d-4644-927a-9c9b122f122a","Type":"ContainerDied","Data":"e948f901034ac36e70079a8d9a37da4905b906e22946fca778dfb603ddc50e86"} Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.299256 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" event={"ID":"9fec464c-ae40-4d5f-8303-ed2c93babb81","Type":"ContainerStarted","Data":"178220fae0bc9c51dc60a3a0b03668f8908ffe282d2d5ecc0e51e04557ad680c"} Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.301112 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.345890 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" podStartSLOduration=2.143762694 podStartE2EDuration="7.345871766s" podCreationTimestamp="2025-10-02 09:45:49 +0000 UTC" firstStartedPulling="2025-10-02 09:45:50.307816385 +0000 UTC m=+1115.664160410" lastFinishedPulling="2025-10-02 09:45:55.509925457 +0000 UTC m=+1120.866269482" observedRunningTime="2025-10-02 09:45:56.340710825 +0000 UTC m=+1121.697054870" watchObservedRunningTime="2025-10-02 09:45:56.345871766 +0000 UTC m=+1121.702215791" Oct 02 09:45:56 crc kubenswrapper[5035]: I1002 09:45:56.972214 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.164344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-combined-ca-bundle\") pod \"1af113d9-450d-4644-927a-9c9b122f122a\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.164408 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-scripts\") pod \"1af113d9-450d-4644-927a-9c9b122f122a\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.164552 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data\") pod \"1af113d9-450d-4644-927a-9c9b122f122a\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.164586 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmz88\" (UniqueName: \"kubernetes.io/projected/1af113d9-450d-4644-927a-9c9b122f122a-kube-api-access-gmz88\") pod \"1af113d9-450d-4644-927a-9c9b122f122a\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.164649 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1af113d9-450d-4644-927a-9c9b122f122a-etc-machine-id\") pod \"1af113d9-450d-4644-927a-9c9b122f122a\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.164688 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data-custom\") pod \"1af113d9-450d-4644-927a-9c9b122f122a\" (UID: \"1af113d9-450d-4644-927a-9c9b122f122a\") " Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.168633 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1af113d9-450d-4644-927a-9c9b122f122a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1af113d9-450d-4644-927a-9c9b122f122a" (UID: "1af113d9-450d-4644-927a-9c9b122f122a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.171448 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af113d9-450d-4644-927a-9c9b122f122a-kube-api-access-gmz88" (OuterVolumeSpecName: "kube-api-access-gmz88") pod "1af113d9-450d-4644-927a-9c9b122f122a" (UID: "1af113d9-450d-4644-927a-9c9b122f122a"). InnerVolumeSpecName "kube-api-access-gmz88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.173119 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1af113d9-450d-4644-927a-9c9b122f122a" (UID: "1af113d9-450d-4644-927a-9c9b122f122a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.173495 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-scripts" (OuterVolumeSpecName: "scripts") pod "1af113d9-450d-4644-927a-9c9b122f122a" (UID: "1af113d9-450d-4644-927a-9c9b122f122a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.260717 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1af113d9-450d-4644-927a-9c9b122f122a" (UID: "1af113d9-450d-4644-927a-9c9b122f122a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.266973 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmz88\" (UniqueName: \"kubernetes.io/projected/1af113d9-450d-4644-927a-9c9b122f122a-kube-api-access-gmz88\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.267003 5035 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1af113d9-450d-4644-927a-9c9b122f122a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.267014 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.267022 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.267032 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.315828 5035 generic.go:334] "Generic (PLEG): container finished" podID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerID="c00b69ba896d36be0e593525a5c9613e9fdb290d350efd1af94d59d38489e81f" exitCode=1 Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.316712 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerDied","Data":"c00b69ba896d36be0e593525a5c9613e9fdb290d350efd1af94d59d38489e81f"} Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.324678 5035 generic.go:334] "Generic (PLEG): container finished" podID="c596018f-cc4e-4dc4-8c37-def6f2b19f94" containerID="fd916db38b6c434f51c2e8bde4ef07b05bd231ff988eaab3e3052ea48e9b33e3" exitCode=0 Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.324907 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7b996fb4b6-gvrhc" event={"ID":"c596018f-cc4e-4dc4-8c37-def6f2b19f94","Type":"ContainerDied","Data":"fd916db38b6c434f51c2e8bde4ef07b05bd231ff988eaab3e3052ea48e9b33e3"} Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.321681 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data" (OuterVolumeSpecName: "config-data") pod "1af113d9-450d-4644-927a-9c9b122f122a" (UID: "1af113d9-450d-4644-927a-9c9b122f122a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.334872 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1af113d9-450d-4644-927a-9c9b122f122a","Type":"ContainerDied","Data":"7959cc01816343aacc49edede813833e82a39e129c9c2bd348f15c7844532d87"} Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.334950 5035 scope.go:117] "RemoveContainer" containerID="83ee2d74a47a74610498a0751f0b680ead41b43110840535555680cee075f32b" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.334965 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.368671 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af113d9-450d-4644-927a-9c9b122f122a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.467956 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.473919 5035 scope.go:117] "RemoveContainer" containerID="e948f901034ac36e70079a8d9a37da4905b906e22946fca778dfb603ddc50e86" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.488936 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.501236 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:57 crc kubenswrapper[5035]: E1002 09:45:57.501648 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="probe" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.501662 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="probe" Oct 02 09:45:57 crc kubenswrapper[5035]: E1002 09:45:57.501687 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="cinder-scheduler" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.501693 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="cinder-scheduler" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.501870 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="cinder-scheduler" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.501887 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af113d9-450d-4644-927a-9c9b122f122a" containerName="probe" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.502839 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.510763 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.514066 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.584712 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-config-data\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.584750 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-995ms\" (UniqueName: \"kubernetes.io/projected/624fbc6b-e52c-4d61-932e-24c8a5686a25-kube-api-access-995ms\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.584833 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/624fbc6b-e52c-4d61-932e-24c8a5686a25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.584896 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.584944 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-scripts\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.584978 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.688376 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/624fbc6b-e52c-4d61-932e-24c8a5686a25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.688475 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.688550 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/624fbc6b-e52c-4d61-932e-24c8a5686a25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.688584 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-scripts\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.689515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.689609 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-config-data\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.689631 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-995ms\" (UniqueName: \"kubernetes.io/projected/624fbc6b-e52c-4d61-932e-24c8a5686a25-kube-api-access-995ms\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.699149 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-scripts\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.710809 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.715131 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-config-data\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.715694 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-995ms\" (UniqueName: \"kubernetes.io/projected/624fbc6b-e52c-4d61-932e-24c8a5686a25-kube-api-access-995ms\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.717853 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/624fbc6b-e52c-4d61-932e-24c8a5686a25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"624fbc6b-e52c-4d61-932e-24c8a5686a25\") " pod="openstack/cinder-scheduler-0" Oct 02 09:45:57 crc kubenswrapper[5035]: I1002 09:45:57.941788 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.178653 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af113d9-450d-4644-927a-9c9b122f122a" path="/var/lib/kubelet/pods/1af113d9-450d-4644-927a-9c9b122f122a/volumes" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.394488 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerStarted","Data":"e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186"} Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.417850 5035 generic.go:334] "Generic (PLEG): container finished" podID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerID="95bff7fafd8b7f1a7bf03d58dab8a569a8b92450b4beec73a560a8c2852071ba" exitCode=0 Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.417976 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6c775994-6t2fx" event={"ID":"91467c75-7c7d-4b20-baad-b168b8eabcbb","Type":"ContainerDied","Data":"95bff7fafd8b7f1a7bf03d58dab8a569a8b92450b4beec73a560a8c2852071ba"} Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.477454 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7b996fb4b6-gvrhc" event={"ID":"c596018f-cc4e-4dc4-8c37-def6f2b19f94","Type":"ContainerStarted","Data":"19b98174de14a32f8c0f47e4a2bc86043aa889865c548fb631bd61fcb307bf39"} Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.477507 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7b996fb4b6-gvrhc" event={"ID":"c596018f-cc4e-4dc4-8c37-def6f2b19f94","Type":"ContainerStarted","Data":"ee470c8cadca33b719987d3662159b942a2343a75fb6ba2f2db78e4b03716db1"} Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.477854 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.549224 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.575691 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-7b996fb4b6-gvrhc" podStartSLOduration=4.652998756 podStartE2EDuration="6.575666104s" podCreationTimestamp="2025-10-02 09:45:52 +0000 UTC" firstStartedPulling="2025-10-02 09:45:53.707854324 +0000 UTC m=+1119.064198349" lastFinishedPulling="2025-10-02 09:45:55.630521672 +0000 UTC m=+1120.986865697" observedRunningTime="2025-10-02 09:45:58.539092904 +0000 UTC m=+1123.895436929" watchObservedRunningTime="2025-10-02 09:45:58.575666104 +0000 UTC m=+1123.932010129" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.701812 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.811189 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.821691 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-combined-ca-bundle\") pod \"91467c75-7c7d-4b20-baad-b168b8eabcbb\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.821885 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data-custom\") pod \"91467c75-7c7d-4b20-baad-b168b8eabcbb\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.821938 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9qzg\" (UniqueName: \"kubernetes.io/projected/91467c75-7c7d-4b20-baad-b168b8eabcbb-kube-api-access-d9qzg\") pod \"91467c75-7c7d-4b20-baad-b168b8eabcbb\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.822009 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data\") pod \"91467c75-7c7d-4b20-baad-b168b8eabcbb\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.822039 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91467c75-7c7d-4b20-baad-b168b8eabcbb-logs\") pod \"91467c75-7c7d-4b20-baad-b168b8eabcbb\" (UID: \"91467c75-7c7d-4b20-baad-b168b8eabcbb\") " Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.827805 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91467c75-7c7d-4b20-baad-b168b8eabcbb-logs" (OuterVolumeSpecName: "logs") pod "91467c75-7c7d-4b20-baad-b168b8eabcbb" (UID: "91467c75-7c7d-4b20-baad-b168b8eabcbb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.852845 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "91467c75-7c7d-4b20-baad-b168b8eabcbb" (UID: "91467c75-7c7d-4b20-baad-b168b8eabcbb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.853864 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91467c75-7c7d-4b20-baad-b168b8eabcbb-kube-api-access-d9qzg" (OuterVolumeSpecName: "kube-api-access-d9qzg") pod "91467c75-7c7d-4b20-baad-b168b8eabcbb" (UID: "91467c75-7c7d-4b20-baad-b168b8eabcbb"). InnerVolumeSpecName "kube-api-access-d9qzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.901648 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data" (OuterVolumeSpecName: "config-data") pod "91467c75-7c7d-4b20-baad-b168b8eabcbb" (UID: "91467c75-7c7d-4b20-baad-b168b8eabcbb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.902225 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91467c75-7c7d-4b20-baad-b168b8eabcbb" (UID: "91467c75-7c7d-4b20-baad-b168b8eabcbb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.924854 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.924897 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91467c75-7c7d-4b20-baad-b168b8eabcbb-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.924907 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.924919 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91467c75-7c7d-4b20-baad-b168b8eabcbb-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:58 crc kubenswrapper[5035]: I1002 09:45:58.924931 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9qzg\" (UniqueName: \"kubernetes.io/projected/91467c75-7c7d-4b20-baad-b168b8eabcbb-kube-api-access-d9qzg\") on node \"crc\" DevicePath \"\"" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.386020 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-5076-account-create-hbnlx"] Oct 02 09:45:59 crc kubenswrapper[5035]: E1002 09:45:59.386473 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api-log" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.386494 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api-log" Oct 02 09:45:59 crc kubenswrapper[5035]: E1002 09:45:59.386510 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.386519 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.389059 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.389100 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api-log" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.389906 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-5076-account-create-hbnlx" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.392062 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-db-secret" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.398073 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-5076-account-create-hbnlx"] Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.512360 5035 generic.go:334] "Generic (PLEG): container finished" podID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerID="e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186" exitCode=0 Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.512402 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerDied","Data":"e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186"} Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.512459 5035 scope.go:117] "RemoveContainer" containerID="c00b69ba896d36be0e593525a5c9613e9fdb290d350efd1af94d59d38489e81f" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.513109 5035 scope.go:117] "RemoveContainer" containerID="c00b69ba896d36be0e593525a5c9613e9fdb290d350efd1af94d59d38489e81f" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.515887 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-f6c775994-6t2fx" event={"ID":"91467c75-7c7d-4b20-baad-b168b8eabcbb","Type":"ContainerDied","Data":"d969915ec6df8b8dbe9c2a65574f09d24938a93579fa72956de4f1f392ca1778"} Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.516085 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-f6c775994-6t2fx" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.535896 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkwsr\" (UniqueName: \"kubernetes.io/projected/b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d-kube-api-access-gkwsr\") pod \"ironic-inspector-5076-account-create-hbnlx\" (UID: \"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d\") " pod="openstack/ironic-inspector-5076-account-create-hbnlx" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.574521 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-f6c775994-6t2fx"] Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.586402 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-f6c775994-6t2fx"] Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.637974 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkwsr\" (UniqueName: \"kubernetes.io/projected/b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d-kube-api-access-gkwsr\") pod \"ironic-inspector-5076-account-create-hbnlx\" (UID: \"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d\") " pod="openstack/ironic-inspector-5076-account-create-hbnlx" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.676290 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkwsr\" (UniqueName: \"kubernetes.io/projected/b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d-kube-api-access-gkwsr\") pod \"ironic-inspector-5076-account-create-hbnlx\" (UID: \"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d\") " pod="openstack/ironic-inspector-5076-account-create-hbnlx" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.709046 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-5076-account-create-hbnlx" Oct 02 09:45:59 crc kubenswrapper[5035]: I1002 09:45:59.963774 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.041032 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.041319 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-central-agent" containerID="cri-o://df35b8f6c564cd8b61a2c0019a9f5f3ef940a2300bd8cc70ac9eeff6bb4288f0" gracePeriod=30 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.041905 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="sg-core" containerID="cri-o://7e2460126f8043a6ac993c056ac4f1abd39edf64d7d97920a5d162c9ebb0c738" gracePeriod=30 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.042088 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="proxy-httpd" containerID="cri-o://969da5960e52410b4f3c237c7b02dfd11c433dfb2001cc4360fba980463adfe0" gracePeriod=30 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.042236 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-notification-agent" containerID="cri-o://2d07daec3490e305d5ebf660b7f949167416a0421bca4ec510dd7ab6235adb3e" gracePeriod=30 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.048987 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.225397 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" path="/var/lib/kubelet/pods/91467c75-7c7d-4b20-baad-b168b8eabcbb/volumes" Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.394173 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.531273 5035 generic.go:334] "Generic (PLEG): container finished" podID="9fec464c-ae40-4d5f-8303-ed2c93babb81" containerID="178220fae0bc9c51dc60a3a0b03668f8908ffe282d2d5ecc0e51e04557ad680c" exitCode=1 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.531344 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" event={"ID":"9fec464c-ae40-4d5f-8303-ed2c93babb81","Type":"ContainerDied","Data":"178220fae0bc9c51dc60a3a0b03668f8908ffe282d2d5ecc0e51e04557ad680c"} Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.532061 5035 scope.go:117] "RemoveContainer" containerID="178220fae0bc9c51dc60a3a0b03668f8908ffe282d2d5ecc0e51e04557ad680c" Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.540127 5035 generic.go:334] "Generic (PLEG): container finished" podID="1275078f-89d8-4873-9ae2-4be7336e399b" containerID="969da5960e52410b4f3c237c7b02dfd11c433dfb2001cc4360fba980463adfe0" exitCode=0 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.540162 5035 generic.go:334] "Generic (PLEG): container finished" podID="1275078f-89d8-4873-9ae2-4be7336e399b" containerID="7e2460126f8043a6ac993c056ac4f1abd39edf64d7d97920a5d162c9ebb0c738" exitCode=2 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.540173 5035 generic.go:334] "Generic (PLEG): container finished" podID="1275078f-89d8-4873-9ae2-4be7336e399b" containerID="df35b8f6c564cd8b61a2c0019a9f5f3ef940a2300bd8cc70ac9eeff6bb4288f0" exitCode=0 Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.540195 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerDied","Data":"969da5960e52410b4f3c237c7b02dfd11c433dfb2001cc4360fba980463adfe0"} Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.540226 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerDied","Data":"7e2460126f8043a6ac993c056ac4f1abd39edf64d7d97920a5d162c9ebb0c738"} Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.540242 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerDied","Data":"df35b8f6c564cd8b61a2c0019a9f5f3ef940a2300bd8cc70ac9eeff6bb4288f0"} Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.851864 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-6hhkr"] Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.853254 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6hhkr" Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.881883 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6hhkr"] Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.956868 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-45c92"] Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.958202 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-45c92" Oct 02 09:46:00 crc kubenswrapper[5035]: I1002 09:46:00.963990 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-45c92"] Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.003752 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2f65\" (UniqueName: \"kubernetes.io/projected/000c2597-bc6a-42d1-96e7-d622bbfe96f4-kube-api-access-d2f65\") pod \"nova-api-db-create-6hhkr\" (UID: \"000c2597-bc6a-42d1-96e7-d622bbfe96f4\") " pod="openstack/nova-api-db-create-6hhkr" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.003838 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff2dd\" (UniqueName: \"kubernetes.io/projected/7a7815e7-389b-44fb-bcaf-feb2c60f8cb5-kube-api-access-ff2dd\") pod \"nova-cell0-db-create-45c92\" (UID: \"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5\") " pod="openstack/nova-cell0-db-create-45c92" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.064849 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-mp9fm"] Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.066502 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mp9fm" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.074441 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-mp9fm"] Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.105014 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6bc467ccf-p6597"] Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.106813 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.106845 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57swk\" (UniqueName: \"kubernetes.io/projected/828d54cf-d9df-4c72-8010-3a39e3e986b7-kube-api-access-57swk\") pod \"nova-cell1-db-create-mp9fm\" (UID: \"828d54cf-d9df-4c72-8010-3a39e3e986b7\") " pod="openstack/nova-cell1-db-create-mp9fm" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.106973 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2f65\" (UniqueName: \"kubernetes.io/projected/000c2597-bc6a-42d1-96e7-d622bbfe96f4-kube-api-access-d2f65\") pod \"nova-api-db-create-6hhkr\" (UID: \"000c2597-bc6a-42d1-96e7-d622bbfe96f4\") " pod="openstack/nova-api-db-create-6hhkr" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.107004 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff2dd\" (UniqueName: \"kubernetes.io/projected/7a7815e7-389b-44fb-bcaf-feb2c60f8cb5-kube-api-access-ff2dd\") pod \"nova-cell0-db-create-45c92\" (UID: \"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5\") " pod="openstack/nova-cell0-db-create-45c92" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.112124 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.112371 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.112492 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.129197 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2f65\" (UniqueName: \"kubernetes.io/projected/000c2597-bc6a-42d1-96e7-d622bbfe96f4-kube-api-access-d2f65\") pod \"nova-api-db-create-6hhkr\" (UID: \"000c2597-bc6a-42d1-96e7-d622bbfe96f4\") " pod="openstack/nova-api-db-create-6hhkr" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.139054 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6bc467ccf-p6597"] Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.143023 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff2dd\" (UniqueName: \"kubernetes.io/projected/7a7815e7-389b-44fb-bcaf-feb2c60f8cb5-kube-api-access-ff2dd\") pod \"nova-cell0-db-create-45c92\" (UID: \"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5\") " pod="openstack/nova-cell0-db-create-45c92" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.174202 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6hhkr" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.209503 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-combined-ca-bundle\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.210436 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-etc-swift\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.210525 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-internal-tls-certs\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.210699 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-config-data\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.210873 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-run-httpd\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.210987 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57swk\" (UniqueName: \"kubernetes.io/projected/828d54cf-d9df-4c72-8010-3a39e3e986b7-kube-api-access-57swk\") pod \"nova-cell1-db-create-mp9fm\" (UID: \"828d54cf-d9df-4c72-8010-3a39e3e986b7\") " pod="openstack/nova-cell1-db-create-mp9fm" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.211039 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc4lz\" (UniqueName: \"kubernetes.io/projected/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-kube-api-access-wc4lz\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.211070 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-public-tls-certs\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.211160 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-log-httpd\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.227889 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57swk\" (UniqueName: \"kubernetes.io/projected/828d54cf-d9df-4c72-8010-3a39e3e986b7-kube-api-access-57swk\") pod \"nova-cell1-db-create-mp9fm\" (UID: \"828d54cf-d9df-4c72-8010-3a39e3e986b7\") " pod="openstack/nova-cell1-db-create-mp9fm" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312610 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-internal-tls-certs\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312671 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-config-data\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312716 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-run-httpd\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312754 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-public-tls-certs\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312770 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc4lz\" (UniqueName: \"kubernetes.io/projected/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-kube-api-access-wc4lz\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312794 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-log-httpd\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312829 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-combined-ca-bundle\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.312863 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-etc-swift\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.313698 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-run-httpd\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.314514 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-log-httpd\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.316988 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-internal-tls-certs\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.317378 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-etc-swift\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.317672 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-config-data\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.317677 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-public-tls-certs\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.322226 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-45c92" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.335232 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc4lz\" (UniqueName: \"kubernetes.io/projected/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-kube-api-access-wc4lz\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.339795 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36-combined-ca-bundle\") pod \"swift-proxy-6bc467ccf-p6597\" (UID: \"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36\") " pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.394970 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mp9fm" Oct 02 09:46:01 crc kubenswrapper[5035]: W1002 09:46:01.427477 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod624fbc6b_e52c_4d61_932e_24c8a5686a25.slice/crio-df3fa7dd908cbf5803fdf666b5198ba8f8a5a8f3be56d61cf93b3cbe7a71e476 WatchSource:0}: Error finding container df3fa7dd908cbf5803fdf666b5198ba8f8a5a8f3be56d61cf93b3cbe7a71e476: Status 404 returned error can't find the container with id df3fa7dd908cbf5803fdf666b5198ba8f8a5a8f3be56d61cf93b3cbe7a71e476 Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.517339 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.554101 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"624fbc6b-e52c-4d61-932e-24c8a5686a25","Type":"ContainerStarted","Data":"df3fa7dd908cbf5803fdf666b5198ba8f8a5a8f3be56d61cf93b3cbe7a71e476"} Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.571577 5035 scope.go:117] "RemoveContainer" containerID="95bff7fafd8b7f1a7bf03d58dab8a569a8b92450b4beec73a560a8c2852071ba" Oct 02 09:46:01 crc kubenswrapper[5035]: E1002 09:46:01.572083 5035 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_init_ironic-7c84bc89d8-vz98z_openstack_bdfd7898-4f4c-4270-9be3-0439621a01a9_0 in pod sandbox 3862e6c597c640e753e6595104433d48f116779719562b0fac4098942b9cb6e8: identifier is not a container" containerID="c00b69ba896d36be0e593525a5c9613e9fdb290d350efd1af94d59d38489e81f" Oct 02 09:46:01 crc kubenswrapper[5035]: E1002 09:46:01.572145 5035 kuberuntime_container.go:896] "Unhandled Error" err="failed to remove pod init container \"init\": rpc error: code = Unknown desc = failed to delete container k8s_init_ironic-7c84bc89d8-vz98z_openstack_bdfd7898-4f4c-4270-9be3-0439621a01a9_0 in pod sandbox 3862e6c597c640e753e6595104433d48f116779719562b0fac4098942b9cb6e8: identifier is not a container; Skipping pod \"ironic-7c84bc89d8-vz98z_openstack(bdfd7898-4f4c-4270-9be3-0439621a01a9)\"" logger="UnhandledError" Oct 02 09:46:01 crc kubenswrapper[5035]: I1002 09:46:01.670579 5035 scope.go:117] "RemoveContainer" containerID="d93282a16895c929b93ba1290144e9be905106614639e0cfc8df3cb492d7061e" Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.246922 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-5076-account-create-hbnlx"] Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.256504 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6hhkr"] Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.335648 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-45c92"] Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.447352 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6f9c95b6cc-5grzm" Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.456081 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-mp9fm"] Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.505201 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58f6c5cff8-5t92g"] Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.505442 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58f6c5cff8-5t92g" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-api" containerID="cri-o://95640f4b076c18d15a63d7cfac95605fe0fce07eb6470486af1f04e53256acac" gracePeriod=30 Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.505898 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-58f6c5cff8-5t92g" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-httpd" containerID="cri-o://173049a936e4990ec816ab9bbdb1fbeca88ec082a9aa7593cdea4d1076f4c304" gracePeriod=30 Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.595664 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6hhkr" event={"ID":"000c2597-bc6a-42d1-96e7-d622bbfe96f4","Type":"ContainerStarted","Data":"952809bd1ae758dcecae925aa8dc95909ac7643b2e54c21449b6943790ecf5af"} Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.598157 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-5076-account-create-hbnlx" event={"ID":"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d","Type":"ContainerStarted","Data":"d189e47261ad13c3e819127f81783b527a503662931b7b3393173df7975ef009"} Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.614810 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-45c92" event={"ID":"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5","Type":"ContainerStarted","Data":"f224c6b61a178f46b216de4de0d24be715614231b35a9d4157b2efea60854546"} Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.630967 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" event={"ID":"9fec464c-ae40-4d5f-8303-ed2c93babb81","Type":"ContainerStarted","Data":"fd91152e4c364836824b8bf271528b90025ce680367cc8168c6b58759b5de242"} Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.632480 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.640867 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerStarted","Data":"5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce"} Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.653704 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-mp9fm" event={"ID":"828d54cf-d9df-4c72-8010-3a39e3e986b7","Type":"ContainerStarted","Data":"57a0fbabbf74544d3933bfc85308eebe0c8c1f173fcb89200fcafcb33441c728"} Oct 02 09:46:02 crc kubenswrapper[5035]: I1002 09:46:02.961625 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6bc467ccf-p6597"] Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.571940 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f6c775994-6t2fx" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.573290 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-f6c775994-6t2fx" podUID="91467c75-7c7d-4b20-baad-b168b8eabcbb" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.669430 5035 generic.go:334] "Generic (PLEG): container finished" podID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerID="f227f16422200928c8e8d4cc28d8e427d7ffa2a3a1d25671b7480411741a3b22" exitCode=1 Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.669512 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerDied","Data":"f227f16422200928c8e8d4cc28d8e427d7ffa2a3a1d25671b7480411741a3b22"} Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.670313 5035 scope.go:117] "RemoveContainer" containerID="f227f16422200928c8e8d4cc28d8e427d7ffa2a3a1d25671b7480411741a3b22" Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.674221 5035 generic.go:334] "Generic (PLEG): container finished" podID="ee399cdc-62d4-409d-8661-5ff797758647" containerID="173049a936e4990ec816ab9bbdb1fbeca88ec082a9aa7593cdea4d1076f4c304" exitCode=0 Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.674347 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58f6c5cff8-5t92g" event={"ID":"ee399cdc-62d4-409d-8661-5ff797758647","Type":"ContainerDied","Data":"173049a936e4990ec816ab9bbdb1fbeca88ec082a9aa7593cdea4d1076f4c304"} Oct 02 09:46:03 crc kubenswrapper[5035]: I1002 09:46:03.677636 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bc467ccf-p6597" event={"ID":"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36","Type":"ContainerStarted","Data":"ae2cf9978437e74d17626f6f684f147f5a371c338b9f577f2ff475d5b5589a63"} Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.689667 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"624fbc6b-e52c-4d61-932e-24c8a5686a25","Type":"ContainerStarted","Data":"95ebfb21514608d401b6d88c650ef6b72f836fdfc91da809fc1ce9185736e1d2"} Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.691059 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-5076-account-create-hbnlx" event={"ID":"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d","Type":"ContainerStarted","Data":"4b7d1a789544379cb5e1fd57aed70c0a43a1d8fec4ddd71aa06c6f6e7989b772"} Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.692277 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6hhkr" event={"ID":"000c2597-bc6a-42d1-96e7-d622bbfe96f4","Type":"ContainerStarted","Data":"ecc30c7d8b0724e3898e58106086877159481cd668ab0423fe8da23f1a477b6e"} Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.693579 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-45c92" event={"ID":"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5","Type":"ContainerStarted","Data":"81ffd6e467018eb41bdb851c2938aa76b37f33fe707a1335da83fa766cc6f684"} Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.696394 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bc467ccf-p6597" event={"ID":"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36","Type":"ContainerStarted","Data":"63de58e005cce847cd3ead6e5ada61e95074e9c895c193fe9e4803e5a8d6d3d2"} Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.697785 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-mp9fm" event={"ID":"828d54cf-d9df-4c72-8010-3a39e3e986b7","Type":"ContainerStarted","Data":"c2429c615c86013bad391a92de956173d184172837bf5ba13809e7e863988ef4"} Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.711969 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-7b996fb4b6-gvrhc" Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.789355 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-7c84bc89d8-vz98z"] Oct 02 09:46:04 crc kubenswrapper[5035]: I1002 09:46:04.975215 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.158:3000/\": dial tcp 10.217.0.158:3000: connect: connection refused" Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.103672 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.104059 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.713019 5035 generic.go:334] "Generic (PLEG): container finished" podID="b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d" containerID="4b7d1a789544379cb5e1fd57aed70c0a43a1d8fec4ddd71aa06c6f6e7989b772" exitCode=0 Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.713076 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-5076-account-create-hbnlx" event={"ID":"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d","Type":"ContainerDied","Data":"4b7d1a789544379cb5e1fd57aed70c0a43a1d8fec4ddd71aa06c6f6e7989b772"} Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.715855 5035 generic.go:334] "Generic (PLEG): container finished" podID="7a7815e7-389b-44fb-bcaf-feb2c60f8cb5" containerID="81ffd6e467018eb41bdb851c2938aa76b37f33fe707a1335da83fa766cc6f684" exitCode=0 Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.715900 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-45c92" event={"ID":"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5","Type":"ContainerDied","Data":"81ffd6e467018eb41bdb851c2938aa76b37f33fe707a1335da83fa766cc6f684"} Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.720452 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6bc467ccf-p6597" event={"ID":"6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36","Type":"ContainerStarted","Data":"203c168f72ac22594489fe034f860e1ec22a0a3375e99701527c31421661dd8c"} Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.720749 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.720778 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.728415 5035 generic.go:334] "Generic (PLEG): container finished" podID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerID="2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433" exitCode=1 Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.728465 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerDied","Data":"2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433"} Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.728496 5035 scope.go:117] "RemoveContainer" containerID="f227f16422200928c8e8d4cc28d8e427d7ffa2a3a1d25671b7480411741a3b22" Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.728860 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ironic-7c84bc89d8-vz98z" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api-log" containerID="cri-o://5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce" gracePeriod=60 Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.750572 5035 generic.go:334] "Generic (PLEG): container finished" podID="1275078f-89d8-4873-9ae2-4be7336e399b" containerID="2d07daec3490e305d5ebf660b7f949167416a0421bca4ec510dd7ab6235adb3e" exitCode=0 Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.750646 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerDied","Data":"2d07daec3490e305d5ebf660b7f949167416a0421bca4ec510dd7ab6235adb3e"} Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.752778 5035 generic.go:334] "Generic (PLEG): container finished" podID="828d54cf-d9df-4c72-8010-3a39e3e986b7" containerID="c2429c615c86013bad391a92de956173d184172837bf5ba13809e7e863988ef4" exitCode=0 Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.752823 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-mp9fm" event={"ID":"828d54cf-d9df-4c72-8010-3a39e3e986b7","Type":"ContainerDied","Data":"c2429c615c86013bad391a92de956173d184172837bf5ba13809e7e863988ef4"} Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.759833 5035 generic.go:334] "Generic (PLEG): container finished" podID="000c2597-bc6a-42d1-96e7-d622bbfe96f4" containerID="ecc30c7d8b0724e3898e58106086877159481cd668ab0423fe8da23f1a477b6e" exitCode=0 Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.759911 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6hhkr" event={"ID":"000c2597-bc6a-42d1-96e7-d622bbfe96f4","Type":"ContainerDied","Data":"ecc30c7d8b0724e3898e58106086877159481cd668ab0423fe8da23f1a477b6e"} Oct 02 09:46:05 crc kubenswrapper[5035]: I1002 09:46:05.818353 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6bc467ccf-p6597" podStartSLOduration=4.818333931 podStartE2EDuration="4.818333931s" podCreationTimestamp="2025-10-02 09:46:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:46:05.816850077 +0000 UTC m=+1131.173194132" watchObservedRunningTime="2025-10-02 09:46:05.818333931 +0000 UTC m=+1131.174677956" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.086927 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.236152 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-log-httpd\") pod \"1275078f-89d8-4873-9ae2-4be7336e399b\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.236254 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-config-data\") pod \"1275078f-89d8-4873-9ae2-4be7336e399b\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.236302 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mct2k\" (UniqueName: \"kubernetes.io/projected/1275078f-89d8-4873-9ae2-4be7336e399b-kube-api-access-mct2k\") pod \"1275078f-89d8-4873-9ae2-4be7336e399b\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.236374 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-combined-ca-bundle\") pod \"1275078f-89d8-4873-9ae2-4be7336e399b\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.236414 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-scripts\") pod \"1275078f-89d8-4873-9ae2-4be7336e399b\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.236473 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-sg-core-conf-yaml\") pod \"1275078f-89d8-4873-9ae2-4be7336e399b\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.236510 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-run-httpd\") pod \"1275078f-89d8-4873-9ae2-4be7336e399b\" (UID: \"1275078f-89d8-4873-9ae2-4be7336e399b\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.237444 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1275078f-89d8-4873-9ae2-4be7336e399b" (UID: "1275078f-89d8-4873-9ae2-4be7336e399b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.241982 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1275078f-89d8-4873-9ae2-4be7336e399b" (UID: "1275078f-89d8-4873-9ae2-4be7336e399b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.253243 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1275078f-89d8-4873-9ae2-4be7336e399b-kube-api-access-mct2k" (OuterVolumeSpecName: "kube-api-access-mct2k") pod "1275078f-89d8-4873-9ae2-4be7336e399b" (UID: "1275078f-89d8-4873-9ae2-4be7336e399b"). InnerVolumeSpecName "kube-api-access-mct2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.253288 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-scripts" (OuterVolumeSpecName: "scripts") pod "1275078f-89d8-4873-9ae2-4be7336e399b" (UID: "1275078f-89d8-4873-9ae2-4be7336e399b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.339032 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.339425 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.339618 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1275078f-89d8-4873-9ae2-4be7336e399b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.339710 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mct2k\" (UniqueName: \"kubernetes.io/projected/1275078f-89d8-4873-9ae2-4be7336e399b-kube-api-access-mct2k\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.364311 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1275078f-89d8-4873-9ae2-4be7336e399b" (UID: "1275078f-89d8-4873-9ae2-4be7336e399b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.423389 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.426245 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1275078f-89d8-4873-9ae2-4be7336e399b" (UID: "1275078f-89d8-4873-9ae2-4be7336e399b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.441664 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.441699 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.448702 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-config-data" (OuterVolumeSpecName: "config-data") pod "1275078f-89d8-4873-9ae2-4be7336e399b" (UID: "1275078f-89d8-4873-9ae2-4be7336e399b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.543389 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-merged\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.543490 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-logs\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.543577 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.543625 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-combined-ca-bundle\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.543650 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-custom\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.543709 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-scripts\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.543730 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/bdfd7898-4f4c-4270-9be3-0439621a01a9-etc-podinfo\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.544326 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-logs" (OuterVolumeSpecName: "logs") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.544651 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.544754 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cqbc\" (UniqueName: \"kubernetes.io/projected/bdfd7898-4f4c-4270-9be3-0439621a01a9-kube-api-access-7cqbc\") pod \"bdfd7898-4f4c-4270-9be3-0439621a01a9\" (UID: \"bdfd7898-4f4c-4270-9be3-0439621a01a9\") " Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.545518 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1275078f-89d8-4873-9ae2-4be7336e399b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.545550 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.545560 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bdfd7898-4f4c-4270-9be3-0439621a01a9-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.554744 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdfd7898-4f4c-4270-9be3-0439621a01a9-kube-api-access-7cqbc" (OuterVolumeSpecName: "kube-api-access-7cqbc") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "kube-api-access-7cqbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.554832 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.555448 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bdfd7898-4f4c-4270-9be3-0439621a01a9-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.555514 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-scripts" (OuterVolumeSpecName: "scripts") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.579780 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data" (OuterVolumeSpecName: "config-data") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.606716 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdfd7898-4f4c-4270-9be3-0439621a01a9" (UID: "bdfd7898-4f4c-4270-9be3-0439621a01a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.648669 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.648699 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.648709 5035 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.648717 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bdfd7898-4f4c-4270-9be3-0439621a01a9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.648726 5035 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/bdfd7898-4f4c-4270-9be3-0439621a01a9-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.648734 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cqbc\" (UniqueName: \"kubernetes.io/projected/bdfd7898-4f4c-4270-9be3-0439621a01a9-kube-api-access-7cqbc\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.777596 5035 generic.go:334] "Generic (PLEG): container finished" podID="9fec464c-ae40-4d5f-8303-ed2c93babb81" containerID="fd91152e4c364836824b8bf271528b90025ce680367cc8168c6b58759b5de242" exitCode=1 Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.777638 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" event={"ID":"9fec464c-ae40-4d5f-8303-ed2c93babb81","Type":"ContainerDied","Data":"fd91152e4c364836824b8bf271528b90025ce680367cc8168c6b58759b5de242"} Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.777934 5035 scope.go:117] "RemoveContainer" containerID="178220fae0bc9c51dc60a3a0b03668f8908ffe282d2d5ecc0e51e04557ad680c" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.778582 5035 scope.go:117] "RemoveContainer" containerID="fd91152e4c364836824b8bf271528b90025ce680367cc8168c6b58759b5de242" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.778869 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-679cc45cbb-lgkg7_openstack(9fec464c-ae40-4d5f-8303-ed2c93babb81)\"" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" podUID="9fec464c-ae40-4d5f-8303-ed2c93babb81" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.783612 5035 generic.go:334] "Generic (PLEG): container finished" podID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerID="5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce" exitCode=143 Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.783709 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-7c84bc89d8-vz98z" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.784118 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerDied","Data":"5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce"} Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.784178 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-7c84bc89d8-vz98z" event={"ID":"bdfd7898-4f4c-4270-9be3-0439621a01a9","Type":"ContainerDied","Data":"3862e6c597c640e753e6595104433d48f116779719562b0fac4098942b9cb6e8"} Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.806021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1275078f-89d8-4873-9ae2-4be7336e399b","Type":"ContainerDied","Data":"5808dffd90cdd0c2dbbda94dda8993709f62e9f250345547b8b486b813b949ad"} Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.806167 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.819585 5035 generic.go:334] "Generic (PLEG): container finished" podID="ee399cdc-62d4-409d-8661-5ff797758647" containerID="95640f4b076c18d15a63d7cfac95605fe0fce07eb6470486af1f04e53256acac" exitCode=0 Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.819641 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58f6c5cff8-5t92g" event={"ID":"ee399cdc-62d4-409d-8661-5ff797758647","Type":"ContainerDied","Data":"95640f4b076c18d15a63d7cfac95605fe0fce07eb6470486af1f04e53256acac"} Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.825858 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-7c84bc89d8-vz98z"] Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.827889 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"624fbc6b-e52c-4d61-932e-24c8a5686a25","Type":"ContainerStarted","Data":"b99c3e3fba3e5fcc902c42467c9c88406766751b31c4b68bf5be3bcde59a2733"} Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.834822 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-7c84bc89d8-vz98z"] Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.873263 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=9.873235619999999 podStartE2EDuration="9.87323562s" podCreationTimestamp="2025-10-02 09:45:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:46:06.852424741 +0000 UTC m=+1132.208768756" watchObservedRunningTime="2025-10-02 09:46:06.87323562 +0000 UTC m=+1132.229579645" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.912829 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.943315 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.969584 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.969973 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.969988 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970001 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="proxy-httpd" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970008 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="proxy-httpd" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970015 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-notification-agent" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970021 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-notification-agent" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970038 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="init" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970046 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="init" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970061 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-central-agent" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970067 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-central-agent" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970077 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="sg-core" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970082 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="sg-core" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970096 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api-log" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970103 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api-log" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970298 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="proxy-httpd" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970315 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-central-agent" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970325 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api-log" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970338 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970350 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="ceilometer-notification-agent" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970358 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970369 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" containerName="sg-core" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970525 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970600 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="ironic-api" Oct 02 09:46:06 crc kubenswrapper[5035]: E1002 09:46:06.970620 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="init" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.970626 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" containerName="init" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.972083 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:06 crc kubenswrapper[5035]: I1002 09:46:06.987621 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.001882 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.002164 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.159804 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.159846 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.159989 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc924\" (UniqueName: \"kubernetes.io/projected/b412869f-797a-4216-997a-3e5adba66a48-kube-api-access-qc924\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.160148 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-config-data\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.160194 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-scripts\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.160233 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-log-httpd\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.160272 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-run-httpd\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.262884 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-config-data\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.262939 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-scripts\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.262967 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-log-httpd\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.262997 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-run-httpd\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.263029 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.263062 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.263171 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc924\" (UniqueName: \"kubernetes.io/projected/b412869f-797a-4216-997a-3e5adba66a48-kube-api-access-qc924\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.271831 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-log-httpd\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.277716 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.280029 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-scripts\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.287061 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-run-httpd\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.293296 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-config-data\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.301476 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc924\" (UniqueName: \"kubernetes.io/projected/b412869f-797a-4216-997a-3e5adba66a48-kube-api-access-qc924\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.319092 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.386043 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:07 crc kubenswrapper[5035]: I1002 09:46:07.942727 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 09:46:08 crc kubenswrapper[5035]: I1002 09:46:08.179758 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1275078f-89d8-4873-9ae2-4be7336e399b" path="/var/lib/kubelet/pods/1275078f-89d8-4873-9ae2-4be7336e399b/volumes" Oct 02 09:46:08 crc kubenswrapper[5035]: I1002 09:46:08.180776 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdfd7898-4f4c-4270-9be3-0439621a01a9" path="/var/lib/kubelet/pods/bdfd7898-4f4c-4270-9be3-0439621a01a9/volumes" Oct 02 09:46:09 crc kubenswrapper[5035]: I1002 09:46:09.693967 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:46:09 crc kubenswrapper[5035]: I1002 09:46:09.694802 5035 scope.go:117] "RemoveContainer" containerID="fd91152e4c364836824b8bf271528b90025ce680367cc8168c6b58759b5de242" Oct 02 09:46:09 crc kubenswrapper[5035]: E1002 09:46:09.695351 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-679cc45cbb-lgkg7_openstack(9fec464c-ae40-4d5f-8303-ed2c93babb81)\"" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" podUID="9fec464c-ae40-4d5f-8303-ed2c93babb81" Oct 02 09:46:10 crc kubenswrapper[5035]: I1002 09:46:10.188052 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:11 crc kubenswrapper[5035]: I1002 09:46:11.523184 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:11 crc kubenswrapper[5035]: I1002 09:46:11.526021 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6bc467ccf-p6597" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.905797 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-45c92" event={"ID":"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5","Type":"ContainerDied","Data":"f224c6b61a178f46b216de4de0d24be715614231b35a9d4157b2efea60854546"} Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.906296 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f224c6b61a178f46b216de4de0d24be715614231b35a9d4157b2efea60854546" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.909853 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-mp9fm" event={"ID":"828d54cf-d9df-4c72-8010-3a39e3e986b7","Type":"ContainerDied","Data":"57a0fbabbf74544d3933bfc85308eebe0c8c1f173fcb89200fcafcb33441c728"} Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.909895 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57a0fbabbf74544d3933bfc85308eebe0c8c1f173fcb89200fcafcb33441c728" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.912375 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-58f6c5cff8-5t92g" event={"ID":"ee399cdc-62d4-409d-8661-5ff797758647","Type":"ContainerDied","Data":"4acdd4bc747f8147c13663bc271cd6dbddbe356e9da7759ff2087a83e5540ec0"} Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.912419 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4acdd4bc747f8147c13663bc271cd6dbddbe356e9da7759ff2087a83e5540ec0" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.917850 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-5076-account-create-hbnlx" event={"ID":"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d","Type":"ContainerDied","Data":"d189e47261ad13c3e819127f81783b527a503662931b7b3393173df7975ef009"} Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.917946 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d189e47261ad13c3e819127f81783b527a503662931b7b3393173df7975ef009" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.922144 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6hhkr" event={"ID":"000c2597-bc6a-42d1-96e7-d622bbfe96f4","Type":"ContainerDied","Data":"952809bd1ae758dcecae925aa8dc95909ac7643b2e54c21449b6943790ecf5af"} Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.922183 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="952809bd1ae758dcecae925aa8dc95909ac7643b2e54c21449b6943790ecf5af" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.936427 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-5076-account-create-hbnlx" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.944899 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-45c92" Oct 02 09:46:12 crc kubenswrapper[5035]: I1002 09:46:12.960930 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6hhkr" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.003079 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mp9fm" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.011682 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.113886 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2f65\" (UniqueName: \"kubernetes.io/projected/000c2597-bc6a-42d1-96e7-d622bbfe96f4-kube-api-access-d2f65\") pod \"000c2597-bc6a-42d1-96e7-d622bbfe96f4\" (UID: \"000c2597-bc6a-42d1-96e7-d622bbfe96f4\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.113974 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57swk\" (UniqueName: \"kubernetes.io/projected/828d54cf-d9df-4c72-8010-3a39e3e986b7-kube-api-access-57swk\") pod \"828d54cf-d9df-4c72-8010-3a39e3e986b7\" (UID: \"828d54cf-d9df-4c72-8010-3a39e3e986b7\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.114010 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkwsr\" (UniqueName: \"kubernetes.io/projected/b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d-kube-api-access-gkwsr\") pod \"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d\" (UID: \"b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.114161 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff2dd\" (UniqueName: \"kubernetes.io/projected/7a7815e7-389b-44fb-bcaf-feb2c60f8cb5-kube-api-access-ff2dd\") pod \"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5\" (UID: \"7a7815e7-389b-44fb-bcaf-feb2c60f8cb5\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.121568 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000c2597-bc6a-42d1-96e7-d622bbfe96f4-kube-api-access-d2f65" (OuterVolumeSpecName: "kube-api-access-d2f65") pod "000c2597-bc6a-42d1-96e7-d622bbfe96f4" (UID: "000c2597-bc6a-42d1-96e7-d622bbfe96f4"). InnerVolumeSpecName "kube-api-access-d2f65". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.124921 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/828d54cf-d9df-4c72-8010-3a39e3e986b7-kube-api-access-57swk" (OuterVolumeSpecName: "kube-api-access-57swk") pod "828d54cf-d9df-4c72-8010-3a39e3e986b7" (UID: "828d54cf-d9df-4c72-8010-3a39e3e986b7"). InnerVolumeSpecName "kube-api-access-57swk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.135867 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d-kube-api-access-gkwsr" (OuterVolumeSpecName: "kube-api-access-gkwsr") pod "b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d" (UID: "b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d"). InnerVolumeSpecName "kube-api-access-gkwsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.135999 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a7815e7-389b-44fb-bcaf-feb2c60f8cb5-kube-api-access-ff2dd" (OuterVolumeSpecName: "kube-api-access-ff2dd") pod "7a7815e7-389b-44fb-bcaf-feb2c60f8cb5" (UID: "7a7815e7-389b-44fb-bcaf-feb2c60f8cb5"). InnerVolumeSpecName "kube-api-access-ff2dd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.216286 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-ovndb-tls-certs\") pod \"ee399cdc-62d4-409d-8661-5ff797758647\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.216329 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-config\") pod \"ee399cdc-62d4-409d-8661-5ff797758647\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.216361 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz72q\" (UniqueName: \"kubernetes.io/projected/ee399cdc-62d4-409d-8661-5ff797758647-kube-api-access-lz72q\") pod \"ee399cdc-62d4-409d-8661-5ff797758647\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.217137 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-combined-ca-bundle\") pod \"ee399cdc-62d4-409d-8661-5ff797758647\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.217181 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-httpd-config\") pod \"ee399cdc-62d4-409d-8661-5ff797758647\" (UID: \"ee399cdc-62d4-409d-8661-5ff797758647\") " Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.217469 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57swk\" (UniqueName: \"kubernetes.io/projected/828d54cf-d9df-4c72-8010-3a39e3e986b7-kube-api-access-57swk\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.217486 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkwsr\" (UniqueName: \"kubernetes.io/projected/b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d-kube-api-access-gkwsr\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.217496 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff2dd\" (UniqueName: \"kubernetes.io/projected/7a7815e7-389b-44fb-bcaf-feb2c60f8cb5-kube-api-access-ff2dd\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.217508 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2f65\" (UniqueName: \"kubernetes.io/projected/000c2597-bc6a-42d1-96e7-d622bbfe96f4-kube-api-access-d2f65\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.219952 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee399cdc-62d4-409d-8661-5ff797758647-kube-api-access-lz72q" (OuterVolumeSpecName: "kube-api-access-lz72q") pod "ee399cdc-62d4-409d-8661-5ff797758647" (UID: "ee399cdc-62d4-409d-8661-5ff797758647"). InnerVolumeSpecName "kube-api-access-lz72q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.227193 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ee399cdc-62d4-409d-8661-5ff797758647" (UID: "ee399cdc-62d4-409d-8661-5ff797758647"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.233522 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.288511 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-config" (OuterVolumeSpecName: "config") pod "ee399cdc-62d4-409d-8661-5ff797758647" (UID: "ee399cdc-62d4-409d-8661-5ff797758647"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.302643 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee399cdc-62d4-409d-8661-5ff797758647" (UID: "ee399cdc-62d4-409d-8661-5ff797758647"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.320315 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.320349 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz72q\" (UniqueName: \"kubernetes.io/projected/ee399cdc-62d4-409d-8661-5ff797758647-kube-api-access-lz72q\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.320362 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.320372 5035 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.334640 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ee399cdc-62d4-409d-8661-5ff797758647" (UID: "ee399cdc-62d4-409d-8661-5ff797758647"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.421662 5035 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee399cdc-62d4-409d-8661-5ff797758647-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.930218 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-5076-account-create-hbnlx" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.930231 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-45c92" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.930241 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6hhkr" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.930269 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-58f6c5cff8-5t92g" Oct 02 09:46:13 crc kubenswrapper[5035]: I1002 09:46:13.930273 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-mp9fm" Oct 02 09:46:14 crc kubenswrapper[5035]: I1002 09:46:14.043599 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-58f6c5cff8-5t92g"] Oct 02 09:46:14 crc kubenswrapper[5035]: I1002 09:46:14.052000 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-58f6c5cff8-5t92g"] Oct 02 09:46:14 crc kubenswrapper[5035]: I1002 09:46:14.174743 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee399cdc-62d4-409d-8661-5ff797758647" path="/var/lib/kubelet/pods/ee399cdc-62d4-409d-8661-5ff797758647/volumes" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.034617 5035 scope.go:117] "RemoveContainer" containerID="2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.093845 5035 scope.go:117] "RemoveContainer" containerID="5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.274798 5035 scope.go:117] "RemoveContainer" containerID="e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.313254 5035 scope.go:117] "RemoveContainer" containerID="2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433" Oct 02 09:46:17 crc kubenswrapper[5035]: E1002 09:46:17.313717 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433\": container with ID starting with 2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433 not found: ID does not exist" containerID="2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.313769 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433"} err="failed to get container status \"2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433\": rpc error: code = NotFound desc = could not find container \"2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433\": container with ID starting with 2d4914af14bd793ef7f1be71d898040bb57272b4d75c31223fca1183d954b433 not found: ID does not exist" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.313795 5035 scope.go:117] "RemoveContainer" containerID="5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce" Oct 02 09:46:17 crc kubenswrapper[5035]: E1002 09:46:17.316162 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce\": container with ID starting with 5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce not found: ID does not exist" containerID="5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.316208 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce"} err="failed to get container status \"5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce\": rpc error: code = NotFound desc = could not find container \"5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce\": container with ID starting with 5d52079627e946f849b1e550608c9915e72d513e86dd4828b4cfd846fb29a9ce not found: ID does not exist" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.316240 5035 scope.go:117] "RemoveContainer" containerID="e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186" Oct 02 09:46:17 crc kubenswrapper[5035]: E1002 09:46:17.317248 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186\": container with ID starting with e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186 not found: ID does not exist" containerID="e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.317276 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186"} err="failed to get container status \"e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186\": rpc error: code = NotFound desc = could not find container \"e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186\": container with ID starting with e0fa2b47e1a9bad45fa7d23e765e8354e8f559d72d4b68a3f1c6888c28c34186 not found: ID does not exist" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.317291 5035 scope.go:117] "RemoveContainer" containerID="969da5960e52410b4f3c237c7b02dfd11c433dfb2001cc4360fba980463adfe0" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.343543 5035 scope.go:117] "RemoveContainer" containerID="7e2460126f8043a6ac993c056ac4f1abd39edf64d7d97920a5d162c9ebb0c738" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.366888 5035 scope.go:117] "RemoveContainer" containerID="2d07daec3490e305d5ebf660b7f949167416a0421bca4ec510dd7ab6235adb3e" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.393808 5035 scope.go:117] "RemoveContainer" containerID="df35b8f6c564cd8b61a2c0019a9f5f3ef940a2300bd8cc70ac9eeff6bb4288f0" Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.548409 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.984982 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"68b71f3f-e46f-4025-9935-e000c44652b5","Type":"ContainerStarted","Data":"e64773293707990e3913637c398629eb99eff63c0b46962c8e3978f3a34a0e40"} Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.992357 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerStarted","Data":"d03fbd025166a00b3a218a4bab54ffc7402aaca54222e6b2038be8a611ed3948"} Oct 02 09:46:17 crc kubenswrapper[5035]: I1002 09:46:17.994917 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerStarted","Data":"a235195dcd74326a780be21bc11ca0a4cacf135d9f7aeeb8336ea2f489cd752c"} Oct 02 09:46:18 crc kubenswrapper[5035]: I1002 09:46:18.023989 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.95485142 podStartE2EDuration="25.023970848s" podCreationTimestamp="2025-10-02 09:45:53 +0000 UTC" firstStartedPulling="2025-10-02 09:45:56.029672832 +0000 UTC m=+1121.386016857" lastFinishedPulling="2025-10-02 09:46:17.09879226 +0000 UTC m=+1142.455136285" observedRunningTime="2025-10-02 09:46:18.02166693 +0000 UTC m=+1143.378010955" watchObservedRunningTime="2025-10-02 09:46:18.023970848 +0000 UTC m=+1143.380314863" Oct 02 09:46:19 crc kubenswrapper[5035]: I1002 09:46:19.005910 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerStarted","Data":"107f5f574515d62a6a941217e6b9b0fd489cc4bbfb28b269672a34c8f89aa5aa"} Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.061294 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1475-account-create-7fvcf"] Oct 02 09:46:21 crc kubenswrapper[5035]: E1002 09:46:21.062073 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-api" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062090 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-api" Oct 02 09:46:21 crc kubenswrapper[5035]: E1002 09:46:21.062110 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a7815e7-389b-44fb-bcaf-feb2c60f8cb5" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062118 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a7815e7-389b-44fb-bcaf-feb2c60f8cb5" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: E1002 09:46:21.062128 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d" containerName="mariadb-account-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062135 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d" containerName="mariadb-account-create" Oct 02 09:46:21 crc kubenswrapper[5035]: E1002 09:46:21.062159 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828d54cf-d9df-4c72-8010-3a39e3e986b7" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062166 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="828d54cf-d9df-4c72-8010-3a39e3e986b7" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: E1002 09:46:21.062195 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-httpd" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062202 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-httpd" Oct 02 09:46:21 crc kubenswrapper[5035]: E1002 09:46:21.062218 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000c2597-bc6a-42d1-96e7-d622bbfe96f4" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062225 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="000c2597-bc6a-42d1-96e7-d622bbfe96f4" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062452 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d" containerName="mariadb-account-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062478 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-httpd" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062498 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="000c2597-bc6a-42d1-96e7-d622bbfe96f4" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062516 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a7815e7-389b-44fb-bcaf-feb2c60f8cb5" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062545 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee399cdc-62d4-409d-8661-5ff797758647" containerName="neutron-api" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.062556 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="828d54cf-d9df-4c72-8010-3a39e3e986b7" containerName="mariadb-database-create" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.063268 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1475-account-create-7fvcf" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.065418 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.069020 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1475-account-create-7fvcf"] Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.182319 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr878\" (UniqueName: \"kubernetes.io/projected/bee75d89-2175-4d59-bd8a-ff46df5eb123-kube-api-access-kr878\") pod \"nova-api-1475-account-create-7fvcf\" (UID: \"bee75d89-2175-4d59-bd8a-ff46df5eb123\") " pod="openstack/nova-api-1475-account-create-7fvcf" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.246269 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-1f39-account-create-2j8xx"] Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.247403 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1f39-account-create-2j8xx" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.251638 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.259803 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1f39-account-create-2j8xx"] Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.311781 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr878\" (UniqueName: \"kubernetes.io/projected/bee75d89-2175-4d59-bd8a-ff46df5eb123-kube-api-access-kr878\") pod \"nova-api-1475-account-create-7fvcf\" (UID: \"bee75d89-2175-4d59-bd8a-ff46df5eb123\") " pod="openstack/nova-api-1475-account-create-7fvcf" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.312150 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8f54\" (UniqueName: \"kubernetes.io/projected/cacd2b45-3487-4e63-848d-466495eb599e-kube-api-access-m8f54\") pod \"nova-cell0-1f39-account-create-2j8xx\" (UID: \"cacd2b45-3487-4e63-848d-466495eb599e\") " pod="openstack/nova-cell0-1f39-account-create-2j8xx" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.333694 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr878\" (UniqueName: \"kubernetes.io/projected/bee75d89-2175-4d59-bd8a-ff46df5eb123-kube-api-access-kr878\") pod \"nova-api-1475-account-create-7fvcf\" (UID: \"bee75d89-2175-4d59-bd8a-ff46df5eb123\") " pod="openstack/nova-api-1475-account-create-7fvcf" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.391486 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1475-account-create-7fvcf" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.414040 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8f54\" (UniqueName: \"kubernetes.io/projected/cacd2b45-3487-4e63-848d-466495eb599e-kube-api-access-m8f54\") pod \"nova-cell0-1f39-account-create-2j8xx\" (UID: \"cacd2b45-3487-4e63-848d-466495eb599e\") " pod="openstack/nova-cell0-1f39-account-create-2j8xx" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.443194 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8f54\" (UniqueName: \"kubernetes.io/projected/cacd2b45-3487-4e63-848d-466495eb599e-kube-api-access-m8f54\") pod \"nova-cell0-1f39-account-create-2j8xx\" (UID: \"cacd2b45-3487-4e63-848d-466495eb599e\") " pod="openstack/nova-cell0-1f39-account-create-2j8xx" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.492067 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-d8be-account-create-gxvzq"] Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.493306 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8be-account-create-gxvzq" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.507201 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.541820 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d8be-account-create-gxvzq"] Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.626858 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1f39-account-create-2j8xx" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.627800 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qssg\" (UniqueName: \"kubernetes.io/projected/f43396b2-6fd0-4736-a8ba-796f27d0d461-kube-api-access-2qssg\") pod \"nova-cell1-d8be-account-create-gxvzq\" (UID: \"f43396b2-6fd0-4736-a8ba-796f27d0d461\") " pod="openstack/nova-cell1-d8be-account-create-gxvzq" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.730682 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qssg\" (UniqueName: \"kubernetes.io/projected/f43396b2-6fd0-4736-a8ba-796f27d0d461-kube-api-access-2qssg\") pod \"nova-cell1-d8be-account-create-gxvzq\" (UID: \"f43396b2-6fd0-4736-a8ba-796f27d0d461\") " pod="openstack/nova-cell1-d8be-account-create-gxvzq" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.753679 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qssg\" (UniqueName: \"kubernetes.io/projected/f43396b2-6fd0-4736-a8ba-796f27d0d461-kube-api-access-2qssg\") pod \"nova-cell1-d8be-account-create-gxvzq\" (UID: \"f43396b2-6fd0-4736-a8ba-796f27d0d461\") " pod="openstack/nova-cell1-d8be-account-create-gxvzq" Oct 02 09:46:21 crc kubenswrapper[5035]: I1002 09:46:21.927806 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8be-account-create-gxvzq" Oct 02 09:46:22 crc kubenswrapper[5035]: I1002 09:46:22.043681 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1475-account-create-7fvcf"] Oct 02 09:46:22 crc kubenswrapper[5035]: W1002 09:46:22.164521 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcacd2b45_3487_4e63_848d_466495eb599e.slice/crio-74669e82e93cb997f885875d2b3993254e468e370a7dadde8cb1fd5db9766659 WatchSource:0}: Error finding container 74669e82e93cb997f885875d2b3993254e468e370a7dadde8cb1fd5db9766659: Status 404 returned error can't find the container with id 74669e82e93cb997f885875d2b3993254e468e370a7dadde8cb1fd5db9766659 Oct 02 09:46:22 crc kubenswrapper[5035]: I1002 09:46:22.177150 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-1f39-account-create-2j8xx"] Oct 02 09:46:22 crc kubenswrapper[5035]: I1002 09:46:22.368954 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-d8be-account-create-gxvzq"] Oct 02 09:46:23 crc kubenswrapper[5035]: I1002 09:46:23.053465 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1475-account-create-7fvcf" event={"ID":"bee75d89-2175-4d59-bd8a-ff46df5eb123","Type":"ContainerStarted","Data":"da845ff7645105043d7d8a4ed6c64e4812f18ead7c7faddf2f545c0a24931916"} Oct 02 09:46:23 crc kubenswrapper[5035]: I1002 09:46:23.055061 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8be-account-create-gxvzq" event={"ID":"f43396b2-6fd0-4736-a8ba-796f27d0d461","Type":"ContainerStarted","Data":"f7e4c52aba55af66392ab7b2d0fbf6b6e341c15e276673250e56aa1da2142782"} Oct 02 09:46:23 crc kubenswrapper[5035]: I1002 09:46:23.056044 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1f39-account-create-2j8xx" event={"ID":"cacd2b45-3487-4e63-848d-466495eb599e","Type":"ContainerStarted","Data":"74669e82e93cb997f885875d2b3993254e468e370a7dadde8cb1fd5db9766659"} Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.163333 5035 scope.go:117] "RemoveContainer" containerID="fd91152e4c364836824b8bf271528b90025ce680367cc8168c6b58759b5de242" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.717622 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-sync-mmkns"] Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.719188 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.722247 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.722345 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.744683 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-mmkns"] Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.897339 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-scripts\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.898626 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.898687 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-combined-ca-bundle\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.898827 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.898961 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-config\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.899020 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qxvt\" (UniqueName: \"kubernetes.io/projected/66817970-09ee-4638-a1df-577234b71172-kube-api-access-8qxvt\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.899081 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/66817970-09ee-4638-a1df-577234b71172-etc-podinfo\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:24 crc kubenswrapper[5035]: I1002 09:46:24.999915 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-scripts\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:24.999982 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.000040 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-combined-ca-bundle\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.000089 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.000134 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-config\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.000168 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qxvt\" (UniqueName: \"kubernetes.io/projected/66817970-09ee-4638-a1df-577234b71172-kube-api-access-8qxvt\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.000211 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/66817970-09ee-4638-a1df-577234b71172-etc-podinfo\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.000940 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.001339 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.009042 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-combined-ca-bundle\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.009094 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-scripts\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.012059 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/66817970-09ee-4638-a1df-577234b71172-etc-podinfo\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.018439 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-config\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.022112 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qxvt\" (UniqueName: \"kubernetes.io/projected/66817970-09ee-4638-a1df-577234b71172-kube-api-access-8qxvt\") pod \"ironic-inspector-db-sync-mmkns\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.042009 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.077405 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerStarted","Data":"f012521ddeacd116610abb8dfb1b381253c2a1aee294068b62bcae96266bce97"} Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.088624 5035 generic.go:334] "Generic (PLEG): container finished" podID="0307b78d-517d-4442-a0d7-1efa3f46098f" containerID="a235195dcd74326a780be21bc11ca0a4cacf135d9f7aeeb8336ea2f489cd752c" exitCode=0 Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.088717 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerDied","Data":"a235195dcd74326a780be21bc11ca0a4cacf135d9f7aeeb8336ea2f489cd752c"} Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.092040 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" event={"ID":"9fec464c-ae40-4d5f-8303-ed2c93babb81","Type":"ContainerStarted","Data":"28e6749f3008c556ad17e5e41798093e4be8fccd1bc20ca35fb4d6eb3565c4ed"} Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.092883 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.538279 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.538641 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:46:25 crc kubenswrapper[5035]: I1002 09:46:25.709870 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-mmkns"] Oct 02 09:46:26 crc kubenswrapper[5035]: I1002 09:46:26.104058 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8be-account-create-gxvzq" event={"ID":"f43396b2-6fd0-4736-a8ba-796f27d0d461","Type":"ContainerStarted","Data":"d7010f1e2875d85efddf261d1615f7a7534deda7f841e7995e0a969330ce9058"} Oct 02 09:46:26 crc kubenswrapper[5035]: I1002 09:46:26.105856 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1f39-account-create-2j8xx" event={"ID":"cacd2b45-3487-4e63-848d-466495eb599e","Type":"ContainerStarted","Data":"108fbb5286e8dabf702a656539736f6cb58c43a311227b9c5b2d0933303d0e57"} Oct 02 09:46:26 crc kubenswrapper[5035]: I1002 09:46:26.107222 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-mmkns" event={"ID":"66817970-09ee-4638-a1df-577234b71172","Type":"ContainerStarted","Data":"d8e81feb1ec21659f3839e53f54026f675ffaef45ce7aec20021dba3fadbaab4"} Oct 02 09:46:26 crc kubenswrapper[5035]: I1002 09:46:26.109014 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1475-account-create-7fvcf" event={"ID":"bee75d89-2175-4d59-bd8a-ff46df5eb123","Type":"ContainerStarted","Data":"2b9b61429e301eff6bedaadffba119544db4e63fa18686a2c34f5143af4c0b0e"} Oct 02 09:46:26 crc kubenswrapper[5035]: I1002 09:46:26.123872 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-1475-account-create-7fvcf" podStartSLOduration=5.123852256 podStartE2EDuration="5.123852256s" podCreationTimestamp="2025-10-02 09:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:46:26.12365633 +0000 UTC m=+1151.480000355" watchObservedRunningTime="2025-10-02 09:46:26.123852256 +0000 UTC m=+1151.480196271" Oct 02 09:46:27 crc kubenswrapper[5035]: I1002 09:46:27.121294 5035 generic.go:334] "Generic (PLEG): container finished" podID="f43396b2-6fd0-4736-a8ba-796f27d0d461" containerID="d7010f1e2875d85efddf261d1615f7a7534deda7f841e7995e0a969330ce9058" exitCode=0 Oct 02 09:46:27 crc kubenswrapper[5035]: I1002 09:46:27.121379 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8be-account-create-gxvzq" event={"ID":"f43396b2-6fd0-4736-a8ba-796f27d0d461","Type":"ContainerDied","Data":"d7010f1e2875d85efddf261d1615f7a7534deda7f841e7995e0a969330ce9058"} Oct 02 09:46:27 crc kubenswrapper[5035]: I1002 09:46:27.163974 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-1f39-account-create-2j8xx" podStartSLOduration=6.163947973 podStartE2EDuration="6.163947973s" podCreationTimestamp="2025-10-02 09:46:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:46:27.152775076 +0000 UTC m=+1152.509119131" watchObservedRunningTime="2025-10-02 09:46:27.163947973 +0000 UTC m=+1152.520291998" Oct 02 09:46:28 crc kubenswrapper[5035]: I1002 09:46:28.136669 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerStarted","Data":"89f49a2d6f2d26d687c0d1832099f71f5d8f2fa654fc38b33a86f31545020e6e"} Oct 02 09:46:28 crc kubenswrapper[5035]: I1002 09:46:28.516426 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8be-account-create-gxvzq" Oct 02 09:46:28 crc kubenswrapper[5035]: I1002 09:46:28.686265 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qssg\" (UniqueName: \"kubernetes.io/projected/f43396b2-6fd0-4736-a8ba-796f27d0d461-kube-api-access-2qssg\") pod \"f43396b2-6fd0-4736-a8ba-796f27d0d461\" (UID: \"f43396b2-6fd0-4736-a8ba-796f27d0d461\") " Oct 02 09:46:28 crc kubenswrapper[5035]: I1002 09:46:28.691197 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43396b2-6fd0-4736-a8ba-796f27d0d461-kube-api-access-2qssg" (OuterVolumeSpecName: "kube-api-access-2qssg") pod "f43396b2-6fd0-4736-a8ba-796f27d0d461" (UID: "f43396b2-6fd0-4736-a8ba-796f27d0d461"). InnerVolumeSpecName "kube-api-access-2qssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:28 crc kubenswrapper[5035]: I1002 09:46:28.788940 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qssg\" (UniqueName: \"kubernetes.io/projected/f43396b2-6fd0-4736-a8ba-796f27d0d461-kube-api-access-2qssg\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.156761 5035 generic.go:334] "Generic (PLEG): container finished" podID="bee75d89-2175-4d59-bd8a-ff46df5eb123" containerID="2b9b61429e301eff6bedaadffba119544db4e63fa18686a2c34f5143af4c0b0e" exitCode=0 Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.156918 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1475-account-create-7fvcf" event={"ID":"bee75d89-2175-4d59-bd8a-ff46df5eb123","Type":"ContainerDied","Data":"2b9b61429e301eff6bedaadffba119544db4e63fa18686a2c34f5143af4c0b0e"} Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.165237 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-d8be-account-create-gxvzq" Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.165268 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-d8be-account-create-gxvzq" event={"ID":"f43396b2-6fd0-4736-a8ba-796f27d0d461","Type":"ContainerDied","Data":"f7e4c52aba55af66392ab7b2d0fbf6b6e341c15e276673250e56aa1da2142782"} Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.165308 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7e4c52aba55af66392ab7b2d0fbf6b6e341c15e276673250e56aa1da2142782" Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.167311 5035 generic.go:334] "Generic (PLEG): container finished" podID="cacd2b45-3487-4e63-848d-466495eb599e" containerID="108fbb5286e8dabf702a656539736f6cb58c43a311227b9c5b2d0933303d0e57" exitCode=0 Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.167376 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1f39-account-create-2j8xx" event={"ID":"cacd2b45-3487-4e63-848d-466495eb599e","Type":"ContainerDied","Data":"108fbb5286e8dabf702a656539736f6cb58c43a311227b9c5b2d0933303d0e57"} Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.171451 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerStarted","Data":"b1e6a4522955dbd93ee431f8023a9bb549f2db99db1e2cff1b686dbc59cf7a6f"} Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.171653 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-central-agent" containerID="cri-o://107f5f574515d62a6a941217e6b9b0fd489cc4bbfb28b269672a34c8f89aa5aa" gracePeriod=30 Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.171743 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-notification-agent" containerID="cri-o://f012521ddeacd116610abb8dfb1b381253c2a1aee294068b62bcae96266bce97" gracePeriod=30 Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.171749 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="proxy-httpd" containerID="cri-o://b1e6a4522955dbd93ee431f8023a9bb549f2db99db1e2cff1b686dbc59cf7a6f" gracePeriod=30 Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.171741 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="sg-core" containerID="cri-o://89f49a2d6f2d26d687c0d1832099f71f5d8f2fa654fc38b33a86f31545020e6e" gracePeriod=30 Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.171820 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.205326 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=12.115840173 podStartE2EDuration="23.20528039s" podCreationTimestamp="2025-10-02 09:46:06 +0000 UTC" firstStartedPulling="2025-10-02 09:46:17.555247854 +0000 UTC m=+1142.911591879" lastFinishedPulling="2025-10-02 09:46:28.644688071 +0000 UTC m=+1154.001032096" observedRunningTime="2025-10-02 09:46:29.197197124 +0000 UTC m=+1154.553541149" watchObservedRunningTime="2025-10-02 09:46:29.20528039 +0000 UTC m=+1154.561624415" Oct 02 09:46:29 crc kubenswrapper[5035]: I1002 09:46:29.736794 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-679cc45cbb-lgkg7" Oct 02 09:46:30 crc kubenswrapper[5035]: I1002 09:46:30.206629 5035 generic.go:334] "Generic (PLEG): container finished" podID="b412869f-797a-4216-997a-3e5adba66a48" containerID="b1e6a4522955dbd93ee431f8023a9bb549f2db99db1e2cff1b686dbc59cf7a6f" exitCode=0 Oct 02 09:46:30 crc kubenswrapper[5035]: I1002 09:46:30.206868 5035 generic.go:334] "Generic (PLEG): container finished" podID="b412869f-797a-4216-997a-3e5adba66a48" containerID="89f49a2d6f2d26d687c0d1832099f71f5d8f2fa654fc38b33a86f31545020e6e" exitCode=2 Oct 02 09:46:30 crc kubenswrapper[5035]: I1002 09:46:30.206875 5035 generic.go:334] "Generic (PLEG): container finished" podID="b412869f-797a-4216-997a-3e5adba66a48" containerID="f012521ddeacd116610abb8dfb1b381253c2a1aee294068b62bcae96266bce97" exitCode=0 Oct 02 09:46:30 crc kubenswrapper[5035]: I1002 09:46:30.207018 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerDied","Data":"b1e6a4522955dbd93ee431f8023a9bb549f2db99db1e2cff1b686dbc59cf7a6f"} Oct 02 09:46:30 crc kubenswrapper[5035]: I1002 09:46:30.207043 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerDied","Data":"89f49a2d6f2d26d687c0d1832099f71f5d8f2fa654fc38b33a86f31545020e6e"} Oct 02 09:46:30 crc kubenswrapper[5035]: I1002 09:46:30.207053 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerDied","Data":"f012521ddeacd116610abb8dfb1b381253c2a1aee294068b62bcae96266bce97"} Oct 02 09:46:31 crc kubenswrapper[5035]: I1002 09:46:31.223076 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-mmkns" event={"ID":"66817970-09ee-4638-a1df-577234b71172","Type":"ContainerStarted","Data":"12beb37c47518f71b410adfb36665834e122b4a709c19a7886babac034f4bc1f"} Oct 02 09:46:31 crc kubenswrapper[5035]: I1002 09:46:31.245195 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-db-sync-mmkns" podStartSLOduration=2.880438104 podStartE2EDuration="7.245172716s" podCreationTimestamp="2025-10-02 09:46:24 +0000 UTC" firstStartedPulling="2025-10-02 09:46:25.728750635 +0000 UTC m=+1151.085094660" lastFinishedPulling="2025-10-02 09:46:30.093485247 +0000 UTC m=+1155.449829272" observedRunningTime="2025-10-02 09:46:31.239922703 +0000 UTC m=+1156.596266728" watchObservedRunningTime="2025-10-02 09:46:31.245172716 +0000 UTC m=+1156.601516741" Oct 02 09:46:33 crc kubenswrapper[5035]: E1002 09:46:33.272112 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb412869f_797a_4216_997a_3e5adba66a48.slice/crio-107f5f574515d62a6a941217e6b9b0fd489cc4bbfb28b269672a34c8f89aa5aa.scope\": RecentStats: unable to find data in memory cache]" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.270692 5035 generic.go:334] "Generic (PLEG): container finished" podID="b412869f-797a-4216-997a-3e5adba66a48" containerID="107f5f574515d62a6a941217e6b9b0fd489cc4bbfb28b269672a34c8f89aa5aa" exitCode=0 Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.270772 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerDied","Data":"107f5f574515d62a6a941217e6b9b0fd489cc4bbfb28b269672a34c8f89aa5aa"} Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.279620 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-1f39-account-create-2j8xx" event={"ID":"cacd2b45-3487-4e63-848d-466495eb599e","Type":"ContainerDied","Data":"74669e82e93cb997f885875d2b3993254e468e370a7dadde8cb1fd5db9766659"} Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.279674 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74669e82e93cb997f885875d2b3993254e468e370a7dadde8cb1fd5db9766659" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.401238 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1f39-account-create-2j8xx" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.408517 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1475-account-create-7fvcf" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.546081 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr878\" (UniqueName: \"kubernetes.io/projected/bee75d89-2175-4d59-bd8a-ff46df5eb123-kube-api-access-kr878\") pod \"bee75d89-2175-4d59-bd8a-ff46df5eb123\" (UID: \"bee75d89-2175-4d59-bd8a-ff46df5eb123\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.546163 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8f54\" (UniqueName: \"kubernetes.io/projected/cacd2b45-3487-4e63-848d-466495eb599e-kube-api-access-m8f54\") pod \"cacd2b45-3487-4e63-848d-466495eb599e\" (UID: \"cacd2b45-3487-4e63-848d-466495eb599e\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.552067 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cacd2b45-3487-4e63-848d-466495eb599e-kube-api-access-m8f54" (OuterVolumeSpecName: "kube-api-access-m8f54") pod "cacd2b45-3487-4e63-848d-466495eb599e" (UID: "cacd2b45-3487-4e63-848d-466495eb599e"). InnerVolumeSpecName "kube-api-access-m8f54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.555685 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bee75d89-2175-4d59-bd8a-ff46df5eb123-kube-api-access-kr878" (OuterVolumeSpecName: "kube-api-access-kr878") pod "bee75d89-2175-4d59-bd8a-ff46df5eb123" (UID: "bee75d89-2175-4d59-bd8a-ff46df5eb123"). InnerVolumeSpecName "kube-api-access-kr878". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.616639 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.648252 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8f54\" (UniqueName: \"kubernetes.io/projected/cacd2b45-3487-4e63-848d-466495eb599e-kube-api-access-m8f54\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.648286 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr878\" (UniqueName: \"kubernetes.io/projected/bee75d89-2175-4d59-bd8a-ff46df5eb123-kube-api-access-kr878\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.749466 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-scripts\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.749938 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.749974 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-sg-core-conf-yaml\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.750081 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-config-data\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.750120 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-log-httpd\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.750139 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-run-httpd\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.750175 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc924\" (UniqueName: \"kubernetes.io/projected/b412869f-797a-4216-997a-3e5adba66a48-kube-api-access-qc924\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.752807 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.752945 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.754818 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-scripts" (OuterVolumeSpecName: "scripts") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.755685 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b412869f-797a-4216-997a-3e5adba66a48-kube-api-access-qc924" (OuterVolumeSpecName: "kube-api-access-qc924") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "kube-api-access-qc924". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.784010 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.853758 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.854785 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle\") pod \"b412869f-797a-4216-997a-3e5adba66a48\" (UID: \"b412869f-797a-4216-997a-3e5adba66a48\") " Oct 02 09:46:34 crc kubenswrapper[5035]: W1002 09:46:34.855271 5035 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b412869f-797a-4216-997a-3e5adba66a48/volumes/kubernetes.io~secret/combined-ca-bundle Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.855296 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.856149 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.856196 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b412869f-797a-4216-997a-3e5adba66a48-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.856210 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc924\" (UniqueName: \"kubernetes.io/projected/b412869f-797a-4216-997a-3e5adba66a48-kube-api-access-qc924\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.856224 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.856235 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.856248 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.883231 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-config-data" (OuterVolumeSpecName: "config-data") pod "b412869f-797a-4216-997a-3e5adba66a48" (UID: "b412869f-797a-4216-997a-3e5adba66a48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:34 crc kubenswrapper[5035]: I1002 09:46:34.958523 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b412869f-797a-4216-997a-3e5adba66a48-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.292523 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1475-account-create-7fvcf" event={"ID":"bee75d89-2175-4d59-bd8a-ff46df5eb123","Type":"ContainerDied","Data":"da845ff7645105043d7d8a4ed6c64e4812f18ead7c7faddf2f545c0a24931916"} Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.292579 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da845ff7645105043d7d8a4ed6c64e4812f18ead7c7faddf2f545c0a24931916" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.293663 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1475-account-create-7fvcf" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.295004 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b412869f-797a-4216-997a-3e5adba66a48","Type":"ContainerDied","Data":"d03fbd025166a00b3a218a4bab54ffc7402aaca54222e6b2038be8a611ed3948"} Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.295040 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.295068 5035 scope.go:117] "RemoveContainer" containerID="b1e6a4522955dbd93ee431f8023a9bb549f2db99db1e2cff1b686dbc59cf7a6f" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.298076 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerStarted","Data":"7665b703df59e2a68f83503c6bbe8a03e984c54b6e9540d89ec79d2dde08de0a"} Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.302735 5035 generic.go:334] "Generic (PLEG): container finished" podID="66817970-09ee-4638-a1df-577234b71172" containerID="12beb37c47518f71b410adfb36665834e122b4a709c19a7886babac034f4bc1f" exitCode=0 Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.303295 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-1f39-account-create-2j8xx" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.302932 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-mmkns" event={"ID":"66817970-09ee-4638-a1df-577234b71172","Type":"ContainerDied","Data":"12beb37c47518f71b410adfb36665834e122b4a709c19a7886babac034f4bc1f"} Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.322628 5035 scope.go:117] "RemoveContainer" containerID="89f49a2d6f2d26d687c0d1832099f71f5d8f2fa654fc38b33a86f31545020e6e" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.348889 5035 scope.go:117] "RemoveContainer" containerID="f012521ddeacd116610abb8dfb1b381253c2a1aee294068b62bcae96266bce97" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.381075 5035 scope.go:117] "RemoveContainer" containerID="107f5f574515d62a6a941217e6b9b0fd489cc4bbfb28b269672a34c8f89aa5aa" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.394835 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.413225 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423204 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:35 crc kubenswrapper[5035]: E1002 09:46:35.423688 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-central-agent" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423711 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-central-agent" Oct 02 09:46:35 crc kubenswrapper[5035]: E1002 09:46:35.423723 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="sg-core" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423729 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="sg-core" Oct 02 09:46:35 crc kubenswrapper[5035]: E1002 09:46:35.423748 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-notification-agent" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423754 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-notification-agent" Oct 02 09:46:35 crc kubenswrapper[5035]: E1002 09:46:35.423764 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43396b2-6fd0-4736-a8ba-796f27d0d461" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423770 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43396b2-6fd0-4736-a8ba-796f27d0d461" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: E1002 09:46:35.423778 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cacd2b45-3487-4e63-848d-466495eb599e" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423784 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cacd2b45-3487-4e63-848d-466495eb599e" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: E1002 09:46:35.423800 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="proxy-httpd" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423805 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="proxy-httpd" Oct 02 09:46:35 crc kubenswrapper[5035]: E1002 09:46:35.423822 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bee75d89-2175-4d59-bd8a-ff46df5eb123" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.423833 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bee75d89-2175-4d59-bd8a-ff46df5eb123" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.424014 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-notification-agent" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.424031 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="cacd2b45-3487-4e63-848d-466495eb599e" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.424048 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bee75d89-2175-4d59-bd8a-ff46df5eb123" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.424064 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="proxy-httpd" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.424080 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="sg-core" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.424096 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43396b2-6fd0-4736-a8ba-796f27d0d461" containerName="mariadb-account-create" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.424109 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b412869f-797a-4216-997a-3e5adba66a48" containerName="ceilometer-central-agent" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.425849 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.431320 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.431553 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.439583 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.569334 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-scripts\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.569392 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.569416 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-run-httpd\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.569478 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-log-httpd\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.569516 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-config-data\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.569563 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9d9f\" (UniqueName: \"kubernetes.io/projected/96e76475-0fec-46d0-8e6c-d346bfb3b456-kube-api-access-r9d9f\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.569628 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.671483 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-log-httpd\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.671589 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-config-data\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.671639 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9d9f\" (UniqueName: \"kubernetes.io/projected/96e76475-0fec-46d0-8e6c-d346bfb3b456-kube-api-access-r9d9f\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.671699 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.671758 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-scripts\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.671793 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.671820 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-run-httpd\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.672187 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-run-httpd\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.673199 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-log-httpd\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.676851 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.677318 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-config-data\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.696704 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.696967 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-scripts\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.701067 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9d9f\" (UniqueName: \"kubernetes.io/projected/96e76475-0fec-46d0-8e6c-d346bfb3b456-kube-api-access-r9d9f\") pod \"ceilometer-0\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " pod="openstack/ceilometer-0" Oct 02 09:46:35 crc kubenswrapper[5035]: I1002 09:46:35.752845 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.176230 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b412869f-797a-4216-997a-3e5adba66a48" path="/var/lib/kubelet/pods/b412869f-797a-4216-997a-3e5adba66a48/volumes" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.348904 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:36 crc kubenswrapper[5035]: W1002 09:46:36.352454 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96e76475_0fec_46d0_8e6c_d346bfb3b456.slice/crio-4fd1436ac0ffe9fc61608dcbf6465179cb3b938b776f06d85d61b3cee6ae53e6 WatchSource:0}: Error finding container 4fd1436ac0ffe9fc61608dcbf6465179cb3b938b776f06d85d61b3cee6ae53e6: Status 404 returned error can't find the container with id 4fd1436ac0ffe9fc61608dcbf6465179cb3b938b776f06d85d61b3cee6ae53e6 Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.531055 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sgtqw"] Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.549976 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.552912 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.553309 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.556213 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b62ln" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.561249 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sgtqw"] Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.699290 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-scripts\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.699349 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvk5m\" (UniqueName: \"kubernetes.io/projected/54ae7cc3-99fc-4485-b5b3-589d64791997-kube-api-access-nvk5m\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.699419 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-config-data\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.699460 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.736229 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.800876 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-config-data\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.800941 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.801022 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-scripts\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.801042 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvk5m\" (UniqueName: \"kubernetes.io/projected/54ae7cc3-99fc-4485-b5b3-589d64791997-kube-api-access-nvk5m\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.805054 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-config-data\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.809411 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-scripts\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.828643 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvk5m\" (UniqueName: \"kubernetes.io/projected/54ae7cc3-99fc-4485-b5b3-589d64791997-kube-api-access-nvk5m\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.833197 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sgtqw\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.873877 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902123 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-scripts\") pod \"66817970-09ee-4638-a1df-577234b71172\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902195 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/66817970-09ee-4638-a1df-577234b71172-etc-podinfo\") pod \"66817970-09ee-4638-a1df-577234b71172\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902291 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic\") pod \"66817970-09ee-4638-a1df-577234b71172\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902361 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-combined-ca-bundle\") pod \"66817970-09ee-4638-a1df-577234b71172\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902418 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"66817970-09ee-4638-a1df-577234b71172\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902496 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-config\") pod \"66817970-09ee-4638-a1df-577234b71172\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902546 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qxvt\" (UniqueName: \"kubernetes.io/projected/66817970-09ee-4638-a1df-577234b71172-kube-api-access-8qxvt\") pod \"66817970-09ee-4638-a1df-577234b71172\" (UID: \"66817970-09ee-4638-a1df-577234b71172\") " Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.902916 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "66817970-09ee-4638-a1df-577234b71172" (UID: "66817970-09ee-4638-a1df-577234b71172"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.903069 5035 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.903737 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "66817970-09ee-4638-a1df-577234b71172" (UID: "66817970-09ee-4638-a1df-577234b71172"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.905631 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/66817970-09ee-4638-a1df-577234b71172-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "66817970-09ee-4638-a1df-577234b71172" (UID: "66817970-09ee-4638-a1df-577234b71172"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.908657 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-scripts" (OuterVolumeSpecName: "scripts") pod "66817970-09ee-4638-a1df-577234b71172" (UID: "66817970-09ee-4638-a1df-577234b71172"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.908861 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66817970-09ee-4638-a1df-577234b71172-kube-api-access-8qxvt" (OuterVolumeSpecName: "kube-api-access-8qxvt") pod "66817970-09ee-4638-a1df-577234b71172" (UID: "66817970-09ee-4638-a1df-577234b71172"). InnerVolumeSpecName "kube-api-access-8qxvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.951585 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66817970-09ee-4638-a1df-577234b71172" (UID: "66817970-09ee-4638-a1df-577234b71172"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:36 crc kubenswrapper[5035]: I1002 09:46:36.957424 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-config" (OuterVolumeSpecName: "config") pod "66817970-09ee-4638-a1df-577234b71172" (UID: "66817970-09ee-4638-a1df-577234b71172"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.004983 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.005014 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qxvt\" (UniqueName: \"kubernetes.io/projected/66817970-09ee-4638-a1df-577234b71172-kube-api-access-8qxvt\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.005026 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.005035 5035 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/66817970-09ee-4638-a1df-577234b71172-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.005045 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66817970-09ee-4638-a1df-577234b71172-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.005053 5035 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/66817970-09ee-4638-a1df-577234b71172-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.323396 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-mmkns" event={"ID":"66817970-09ee-4638-a1df-577234b71172","Type":"ContainerDied","Data":"d8e81feb1ec21659f3839e53f54026f675ffaef45ce7aec20021dba3fadbaab4"} Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.323430 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8e81feb1ec21659f3839e53f54026f675ffaef45ce7aec20021dba3fadbaab4" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.323482 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-mmkns" Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.330738 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerStarted","Data":"19fae54dd1bd848b4159124317b70000e34344a77f8bb966d012b4464d769df5"} Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.330792 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerStarted","Data":"4fd1436ac0ffe9fc61608dcbf6465179cb3b938b776f06d85d61b3cee6ae53e6"} Oct 02 09:46:37 crc kubenswrapper[5035]: W1002 09:46:37.435574 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54ae7cc3_99fc_4485_b5b3_589d64791997.slice/crio-7ca3b7eb4081062e497241b3ceeecf35c99e45101ab0e95e4dc3d2ad70863db1 WatchSource:0}: Error finding container 7ca3b7eb4081062e497241b3ceeecf35c99e45101ab0e95e4dc3d2ad70863db1: Status 404 returned error can't find the container with id 7ca3b7eb4081062e497241b3ceeecf35c99e45101ab0e95e4dc3d2ad70863db1 Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.444864 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sgtqw"] Oct 02 09:46:37 crc kubenswrapper[5035]: I1002 09:46:37.852346 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:38 crc kubenswrapper[5035]: I1002 09:46:38.343397 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" event={"ID":"54ae7cc3-99fc-4485-b5b3-589d64791997","Type":"ContainerStarted","Data":"7ca3b7eb4081062e497241b3ceeecf35c99e45101ab0e95e4dc3d2ad70863db1"} Oct 02 09:46:38 crc kubenswrapper[5035]: I1002 09:46:38.344873 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerStarted","Data":"c3fea9636911d0bdcc0510bd44047092d0c3d3bb1f8594c1003bdafbcd3fb3f2"} Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.013073 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:39 crc kubenswrapper[5035]: E1002 09:46:39.016331 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66817970-09ee-4638-a1df-577234b71172" containerName="ironic-inspector-db-sync" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.016609 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="66817970-09ee-4638-a1df-577234b71172" containerName="ironic-inspector-db-sync" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.016956 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="66817970-09ee-4638-a1df-577234b71172" containerName="ironic-inspector-db-sync" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.020307 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.035848 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.036100 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.048897 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.152021 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.153255 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.153457 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1417b3e1-74ef-46eb-ab3c-c98f708c3688-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.153611 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg8cg\" (UniqueName: \"kubernetes.io/projected/1417b3e1-74ef-46eb-ab3c-c98f708c3688-kube-api-access-tg8cg\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.153758 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-config\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.153960 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.154107 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-scripts\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.255293 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.256367 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.257131 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1417b3e1-74ef-46eb-ab3c-c98f708c3688-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.257833 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg8cg\" (UniqueName: \"kubernetes.io/projected/1417b3e1-74ef-46eb-ab3c-c98f708c3688-kube-api-access-tg8cg\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.257987 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-config\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.258232 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.258423 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-scripts\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.256849 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.256095 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.262347 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1417b3e1-74ef-46eb-ab3c-c98f708c3688-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.262593 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-scripts\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.263855 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-config\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.272105 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.281140 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg8cg\" (UniqueName: \"kubernetes.io/projected/1417b3e1-74ef-46eb-ab3c-c98f708c3688-kube-api-access-tg8cg\") pod \"ironic-inspector-0\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.370683 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.371643 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerStarted","Data":"3e2a2d08fddfab03d09c2ac4491ed93b33809cbfe5f53c15ed51531ecf4c3cfa"} Oct 02 09:46:39 crc kubenswrapper[5035]: I1002 09:46:39.803089 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.383378 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerStarted","Data":"6af8c80800b309d2da419734305b2d58faf447ba1d49c041eb7614b6652f54e1"} Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.383746 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-central-agent" containerID="cri-o://19fae54dd1bd848b4159124317b70000e34344a77f8bb966d012b4464d769df5" gracePeriod=30 Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.384053 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="sg-core" containerID="cri-o://3e2a2d08fddfab03d09c2ac4491ed93b33809cbfe5f53c15ed51531ecf4c3cfa" gracePeriod=30 Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.384230 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.384065 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-notification-agent" containerID="cri-o://c3fea9636911d0bdcc0510bd44047092d0c3d3bb1f8594c1003bdafbcd3fb3f2" gracePeriod=30 Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.384099 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="proxy-httpd" containerID="cri-o://6af8c80800b309d2da419734305b2d58faf447ba1d49c041eb7614b6652f54e1" gracePeriod=30 Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.390064 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"1417b3e1-74ef-46eb-ab3c-c98f708c3688","Type":"ContainerStarted","Data":"e8426c2d31ab59673f88da5dd477a07692db4dd977d3b6ec6f2139cff7182a98"} Oct 02 09:46:40 crc kubenswrapper[5035]: I1002 09:46:40.411620 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.099616768 podStartE2EDuration="5.411604373s" podCreationTimestamp="2025-10-02 09:46:35 +0000 UTC" firstStartedPulling="2025-10-02 09:46:36.356453952 +0000 UTC m=+1161.712797977" lastFinishedPulling="2025-10-02 09:46:39.668441557 +0000 UTC m=+1165.024785582" observedRunningTime="2025-10-02 09:46:40.408961426 +0000 UTC m=+1165.765305451" watchObservedRunningTime="2025-10-02 09:46:40.411604373 +0000 UTC m=+1165.767948398" Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.178113 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.180113 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-log" containerID="cri-o://f2c89a8d7b3ee8367edfb6af778e14bc3505cf972e0f8764ea0d989c7b3e5d34" gracePeriod=30 Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.180248 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-httpd" containerID="cri-o://6ea426aa25725e5a82f3a3f4772279450441598ebdc4296ba755846b29b4fd6a" gracePeriod=30 Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.404337 5035 generic.go:334] "Generic (PLEG): container finished" podID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerID="f2c89a8d7b3ee8367edfb6af778e14bc3505cf972e0f8764ea0d989c7b3e5d34" exitCode=143 Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.404410 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442","Type":"ContainerDied","Data":"f2c89a8d7b3ee8367edfb6af778e14bc3505cf972e0f8764ea0d989c7b3e5d34"} Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.408019 5035 generic.go:334] "Generic (PLEG): container finished" podID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerID="6af8c80800b309d2da419734305b2d58faf447ba1d49c041eb7614b6652f54e1" exitCode=0 Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.408052 5035 generic.go:334] "Generic (PLEG): container finished" podID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerID="3e2a2d08fddfab03d09c2ac4491ed93b33809cbfe5f53c15ed51531ecf4c3cfa" exitCode=2 Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.408063 5035 generic.go:334] "Generic (PLEG): container finished" podID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerID="c3fea9636911d0bdcc0510bd44047092d0c3d3bb1f8594c1003bdafbcd3fb3f2" exitCode=0 Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.408112 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerDied","Data":"6af8c80800b309d2da419734305b2d58faf447ba1d49c041eb7614b6652f54e1"} Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.408142 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerDied","Data":"3e2a2d08fddfab03d09c2ac4491ed93b33809cbfe5f53c15ed51531ecf4c3cfa"} Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.408154 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerDied","Data":"c3fea9636911d0bdcc0510bd44047092d0c3d3bb1f8594c1003bdafbcd3fb3f2"} Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.413065 5035 generic.go:334] "Generic (PLEG): container finished" podID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" containerID="ee9375d91aab70bed1379563b9054b67667a95a24627db26e986c25bac1cad5e" exitCode=0 Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.413112 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"1417b3e1-74ef-46eb-ab3c-c98f708c3688","Type":"ContainerDied","Data":"ee9375d91aab70bed1379563b9054b67667a95a24627db26e986c25bac1cad5e"} Oct 02 09:46:41 crc kubenswrapper[5035]: I1002 09:46:41.992127 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:42 crc kubenswrapper[5035]: I1002 09:46:42.440822 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"1417b3e1-74ef-46eb-ab3c-c98f708c3688","Type":"ContainerStarted","Data":"622091f5b215247208af56a8c6910d3577bc3c313113973b8976dbbd8fb501a4"} Oct 02 09:46:42 crc kubenswrapper[5035]: I1002 09:46:42.487543 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:46:42 crc kubenswrapper[5035]: I1002 09:46:42.487805 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-log" containerID="cri-o://c602845ee5438d0817f63a5900578511888dcf10b25a37336986d1f8931e2f29" gracePeriod=30 Oct 02 09:46:42 crc kubenswrapper[5035]: I1002 09:46:42.488032 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-httpd" containerID="cri-o://804befe089b471e947159d43e917b186670d15949200f28417981b3bfffae808" gracePeriod=30 Oct 02 09:46:43 crc kubenswrapper[5035]: I1002 09:46:43.473786 5035 generic.go:334] "Generic (PLEG): container finished" podID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerID="c602845ee5438d0817f63a5900578511888dcf10b25a37336986d1f8931e2f29" exitCode=143 Oct 02 09:46:43 crc kubenswrapper[5035]: I1002 09:46:43.474112 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f42b1109-c1b4-4a08-a8d3-03cc50774a90","Type":"ContainerDied","Data":"c602845ee5438d0817f63a5900578511888dcf10b25a37336986d1f8931e2f29"} Oct 02 09:46:43 crc kubenswrapper[5035]: I1002 09:46:43.477989 5035 generic.go:334] "Generic (PLEG): container finished" podID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" containerID="622091f5b215247208af56a8c6910d3577bc3c313113973b8976dbbd8fb501a4" exitCode=0 Oct 02 09:46:43 crc kubenswrapper[5035]: I1002 09:46:43.478022 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"1417b3e1-74ef-46eb-ab3c-c98f708c3688","Type":"ContainerDied","Data":"622091f5b215247208af56a8c6910d3577bc3c313113973b8976dbbd8fb501a4"} Oct 02 09:46:44 crc kubenswrapper[5035]: I1002 09:46:44.488937 5035 generic.go:334] "Generic (PLEG): container finished" podID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerID="6ea426aa25725e5a82f3a3f4772279450441598ebdc4296ba755846b29b4fd6a" exitCode=0 Oct 02 09:46:44 crc kubenswrapper[5035]: I1002 09:46:44.489021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442","Type":"ContainerDied","Data":"6ea426aa25725e5a82f3a3f4772279450441598ebdc4296ba755846b29b4fd6a"} Oct 02 09:46:45 crc kubenswrapper[5035]: I1002 09:46:45.639292 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:45196->10.217.0.150:9292: read: connection reset by peer" Oct 02 09:46:45 crc kubenswrapper[5035]: I1002 09:46:45.639320 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:45186->10.217.0.150:9292: read: connection reset by peer" Oct 02 09:46:46 crc kubenswrapper[5035]: I1002 09:46:46.507687 5035 generic.go:334] "Generic (PLEG): container finished" podID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerID="804befe089b471e947159d43e917b186670d15949200f28417981b3bfffae808" exitCode=0 Oct 02 09:46:46 crc kubenswrapper[5035]: I1002 09:46:46.507721 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f42b1109-c1b4-4a08-a8d3-03cc50774a90","Type":"ContainerDied","Data":"804befe089b471e947159d43e917b186670d15949200f28417981b3bfffae808"} Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.540595 5035 generic.go:334] "Generic (PLEG): container finished" podID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerID="19fae54dd1bd848b4159124317b70000e34344a77f8bb966d012b4464d769df5" exitCode=0 Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.540998 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerDied","Data":"19fae54dd1bd848b4159124317b70000e34344a77f8bb966d012b4464d769df5"} Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.549582 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"1417b3e1-74ef-46eb-ab3c-c98f708c3688","Type":"ContainerDied","Data":"e8426c2d31ab59673f88da5dd477a07692db4dd977d3b6ec6f2139cff7182a98"} Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.549645 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8426c2d31ab59673f88da5dd477a07692db4dd977d3b6ec6f2139cff7182a98" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.580044 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.664863 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg8cg\" (UniqueName: \"kubernetes.io/projected/1417b3e1-74ef-46eb-ab3c-c98f708c3688-kube-api-access-tg8cg\") pod \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.665273 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-combined-ca-bundle\") pod \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.665459 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-scripts\") pod \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.665519 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.665583 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1417b3e1-74ef-46eb-ab3c-c98f708c3688-etc-podinfo\") pod \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.665613 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic\") pod \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.665689 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-config\") pod \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\" (UID: \"1417b3e1-74ef-46eb-ab3c-c98f708c3688\") " Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.666638 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "1417b3e1-74ef-46eb-ab3c-c98f708c3688" (UID: "1417b3e1-74ef-46eb-ab3c-c98f708c3688"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.669044 5035 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.679484 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "1417b3e1-74ef-46eb-ab3c-c98f708c3688" (UID: "1417b3e1-74ef-46eb-ab3c-c98f708c3688"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.682566 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1417b3e1-74ef-46eb-ab3c-c98f708c3688-kube-api-access-tg8cg" (OuterVolumeSpecName: "kube-api-access-tg8cg") pod "1417b3e1-74ef-46eb-ab3c-c98f708c3688" (UID: "1417b3e1-74ef-46eb-ab3c-c98f708c3688"). InnerVolumeSpecName "kube-api-access-tg8cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.687699 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-config" (OuterVolumeSpecName: "config") pod "1417b3e1-74ef-46eb-ab3c-c98f708c3688" (UID: "1417b3e1-74ef-46eb-ab3c-c98f708c3688"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.697725 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1417b3e1-74ef-46eb-ab3c-c98f708c3688-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "1417b3e1-74ef-46eb-ab3c-c98f708c3688" (UID: "1417b3e1-74ef-46eb-ab3c-c98f708c3688"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.700670 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-scripts" (OuterVolumeSpecName: "scripts") pod "1417b3e1-74ef-46eb-ab3c-c98f708c3688" (UID: "1417b3e1-74ef-46eb-ab3c-c98f708c3688"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.778106 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.778144 5035 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1417b3e1-74ef-46eb-ab3c-c98f708c3688-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.778158 5035 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1417b3e1-74ef-46eb-ab3c-c98f708c3688-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.778167 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.778176 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg8cg\" (UniqueName: \"kubernetes.io/projected/1417b3e1-74ef-46eb-ab3c-c98f708c3688-kube-api-access-tg8cg\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.856120 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1417b3e1-74ef-46eb-ab3c-c98f708c3688" (UID: "1417b3e1-74ef-46eb-ab3c-c98f708c3688"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:48 crc kubenswrapper[5035]: I1002 09:46:48.880323 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1417b3e1-74ef-46eb-ab3c-c98f708c3688-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.562558 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442","Type":"ContainerDied","Data":"0fcc82bf265838d8a9ec6e6d07cb82e193e3c0fceb522d410599cdf18c9f7ad8"} Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.562847 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fcc82bf265838d8a9ec6e6d07cb82e193e3c0fceb522d410599cdf18c9f7ad8" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.565897 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"96e76475-0fec-46d0-8e6c-d346bfb3b456","Type":"ContainerDied","Data":"4fd1436ac0ffe9fc61608dcbf6465179cb3b938b776f06d85d61b3cee6ae53e6"} Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.565929 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fd1436ac0ffe9fc61608dcbf6465179cb3b938b776f06d85d61b3cee6ae53e6" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.569661 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.570081 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f42b1109-c1b4-4a08-a8d3-03cc50774a90","Type":"ContainerDied","Data":"806d31a90204ebd1fcda3c8f699567c2cd0d3e8883ec2fabcdfeef0f31a68aee"} Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.570107 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="806d31a90204ebd1fcda3c8f699567c2cd0d3e8883ec2fabcdfeef0f31a68aee" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.638183 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.647947 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.656576 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.692365 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.697982 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732106 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-config-data\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732162 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-sg-core-conf-yaml\") pod \"96e76475-0fec-46d0-8e6c-d346bfb3b456\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732231 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-combined-ca-bundle\") pod \"96e76475-0fec-46d0-8e6c-d346bfb3b456\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732262 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-config-data\") pod \"96e76475-0fec-46d0-8e6c-d346bfb3b456\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732288 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-scripts\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732326 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-httpd-run\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732351 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-combined-ca-bundle\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732394 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69wj2\" (UniqueName: \"kubernetes.io/projected/f42b1109-c1b4-4a08-a8d3-03cc50774a90-kube-api-access-69wj2\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732441 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-combined-ca-bundle\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732498 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s9s2\" (UniqueName: \"kubernetes.io/projected/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-kube-api-access-5s9s2\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732555 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-scripts\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732581 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-run-httpd\") pod \"96e76475-0fec-46d0-8e6c-d346bfb3b456\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732623 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-log-httpd\") pod \"96e76475-0fec-46d0-8e6c-d346bfb3b456\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732656 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-internal-tls-certs\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732690 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9d9f\" (UniqueName: \"kubernetes.io/projected/96e76475-0fec-46d0-8e6c-d346bfb3b456-kube-api-access-r9d9f\") pod \"96e76475-0fec-46d0-8e6c-d346bfb3b456\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732732 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732776 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-httpd-run\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732803 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-logs\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.732848 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.747739 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e76475-0fec-46d0-8e6c-d346bfb3b456-kube-api-access-r9d9f" (OuterVolumeSpecName: "kube-api-access-r9d9f") pod "96e76475-0fec-46d0-8e6c-d346bfb3b456" (UID: "96e76475-0fec-46d0-8e6c-d346bfb3b456"). InnerVolumeSpecName "kube-api-access-r9d9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.759138 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "96e76475-0fec-46d0-8e6c-d346bfb3b456" (UID: "96e76475-0fec-46d0-8e6c-d346bfb3b456"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.759392 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "96e76475-0fec-46d0-8e6c-d346bfb3b456" (UID: "96e76475-0fec-46d0-8e6c-d346bfb3b456"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.760726 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.760982 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.761244 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-logs" (OuterVolumeSpecName: "logs") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787169 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787607 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787623 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787631 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-central-agent" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787638 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-central-agent" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787651 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" containerName="ironic-python-agent-init" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787656 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" containerName="ironic-python-agent-init" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787672 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-log" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787680 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-log" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787689 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="proxy-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787694 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="proxy-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787705 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-log" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787710 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-log" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787720 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" containerName="inspector-pxe-init" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787726 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" containerName="inspector-pxe-init" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787738 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="sg-core" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787744 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="sg-core" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787761 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-notification-agent" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787768 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-notification-agent" Oct 02 09:46:49 crc kubenswrapper[5035]: E1002 09:46:49.787783 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787788 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787940 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787953 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="sg-core" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787961 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-notification-agent" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787971 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-log" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787980 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="proxy-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787992 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" containerName="glance-log" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.787999 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" containerName="ceilometer-central-agent" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.788009 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" containerName="inspector-pxe-init" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.788023 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" containerName="glance-httpd" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.790342 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.800408 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-scripts" (OuterVolumeSpecName: "scripts") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.800688 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.800769 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-internal-svc" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.800867 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.801043 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.801096 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-public-svc" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.801215 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.814946 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-kube-api-access-5s9s2" (OuterVolumeSpecName: "kube-api-access-5s9s2") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "kube-api-access-5s9s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.815059 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-scripts" (OuterVolumeSpecName: "scripts") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.828801 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f42b1109-c1b4-4a08-a8d3-03cc50774a90-kube-api-access-69wj2" (OuterVolumeSpecName: "kube-api-access-69wj2") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "kube-api-access-69wj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.834764 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-public-tls-certs\") pod \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\" (UID: \"cc55ee7d-a2e9-4bc7-a86d-9d4d58356442\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.834814 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-config-data\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.834852 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-scripts\") pod \"96e76475-0fec-46d0-8e6c-d346bfb3b456\" (UID: \"96e76475-0fec-46d0-8e6c-d346bfb3b456\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.834893 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-logs\") pod \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\" (UID: \"f42b1109-c1b4-4a08-a8d3-03cc50774a90\") " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.841108 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvcnn\" (UniqueName: \"kubernetes.io/projected/dd4355e8-6dd5-4937-969a-bad47330d25b-kube-api-access-fvcnn\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.841238 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-config\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.841309 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.841431 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/dd4355e8-6dd5-4937-969a-bad47330d25b-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.841479 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/dd4355e8-6dd5-4937-969a-bad47330d25b-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.844260 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-logs" (OuterVolumeSpecName: "logs") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.847740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-scripts\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.847887 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.847951 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.847987 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/dd4355e8-6dd5-4937-969a-bad47330d25b-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.849318 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.854420 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.854522 5035 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f42b1109-c1b4-4a08-a8d3-03cc50774a90-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.854653 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69wj2\" (UniqueName: \"kubernetes.io/projected/f42b1109-c1b4-4a08-a8d3-03cc50774a90-kube-api-access-69wj2\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.854670 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s9s2\" (UniqueName: \"kubernetes.io/projected/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-kube-api-access-5s9s2\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855338 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855358 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855373 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/96e76475-0fec-46d0-8e6c-d346bfb3b456-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855384 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9d9f\" (UniqueName: \"kubernetes.io/projected/96e76475-0fec-46d0-8e6c-d346bfb3b456-kube-api-access-r9d9f\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855413 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855426 5035 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855496 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.855547 5035 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.880959 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.907745 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-scripts" (OuterVolumeSpecName: "scripts") pod "96e76475-0fec-46d0-8e6c-d346bfb3b456" (UID: "96e76475-0fec-46d0-8e6c-d346bfb3b456"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.928697 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "96e76475-0fec-46d0-8e6c-d346bfb3b456" (UID: "96e76475-0fec-46d0-8e6c-d346bfb3b456"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.937639 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.956893 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.956951 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.956977 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/dd4355e8-6dd5-4937-969a-bad47330d25b-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957005 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvcnn\" (UniqueName: \"kubernetes.io/projected/dd4355e8-6dd5-4937-969a-bad47330d25b-kube-api-access-fvcnn\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957030 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-config\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957070 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957123 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/dd4355e8-6dd5-4937-969a-bad47330d25b-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957151 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/dd4355e8-6dd5-4937-969a-bad47330d25b-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957176 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-scripts\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957221 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957232 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957241 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.957481 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/dd4355e8-6dd5-4937-969a-bad47330d25b-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.962436 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.967910 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/dd4355e8-6dd5-4937-969a-bad47330d25b-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.972289 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.989366 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-scripts\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.989475 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.994559 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/dd4355e8-6dd5-4937-969a-bad47330d25b-config\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.996456 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.996653 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:49 crc kubenswrapper[5035]: I1002 09:46:49.998223 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/dd4355e8-6dd5-4937-969a-bad47330d25b-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.006182 5035 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.007795 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvcnn\" (UniqueName: \"kubernetes.io/projected/dd4355e8-6dd5-4937-969a-bad47330d25b-kube-api-access-fvcnn\") pod \"ironic-inspector-0\" (UID: \"dd4355e8-6dd5-4937-969a-bad47330d25b\") " pod="openstack/ironic-inspector-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.050509 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-config-data" (OuterVolumeSpecName: "config-data") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.058937 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.058969 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.058979 5035 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.058989 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.063151 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.065869 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f42b1109-c1b4-4a08-a8d3-03cc50774a90" (UID: "f42b1109-c1b4-4a08-a8d3-03cc50774a90"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.070448 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-config-data" (OuterVolumeSpecName: "config-data") pod "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" (UID: "cc55ee7d-a2e9-4bc7-a86d-9d4d58356442"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.088455 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96e76475-0fec-46d0-8e6c-d346bfb3b456" (UID: "96e76475-0fec-46d0-8e6c-d346bfb3b456"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.117658 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-config-data" (OuterVolumeSpecName: "config-data") pod "96e76475-0fec-46d0-8e6c-d346bfb3b456" (UID: "96e76475-0fec-46d0-8e6c-d346bfb3b456"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.160045 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.160307 5035 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f42b1109-c1b4-4a08-a8d3-03cc50774a90-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.160399 5035 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.160485 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.160592 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e76475-0fec-46d0-8e6c-d346bfb3b456-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.180931 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.186092 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1417b3e1-74ef-46eb-ab3c-c98f708c3688" path="/var/lib/kubelet/pods/1417b3e1-74ef-46eb-ab3c-c98f708c3688/volumes" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.580729 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" event={"ID":"54ae7cc3-99fc-4485-b5b3-589d64791997","Type":"ContainerStarted","Data":"7f617c73dfbdcae4fe1152dd876e3ad4da1920b515ac23133acb5a9b3270619c"} Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.580788 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.580808 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.580864 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.612800 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" podStartSLOduration=2.38375478 podStartE2EDuration="14.612784081s" podCreationTimestamp="2025-10-02 09:46:36 +0000 UTC" firstStartedPulling="2025-10-02 09:46:37.438945868 +0000 UTC m=+1162.795289893" lastFinishedPulling="2025-10-02 09:46:49.667975179 +0000 UTC m=+1175.024319194" observedRunningTime="2025-10-02 09:46:50.608737502 +0000 UTC m=+1175.965081547" watchObservedRunningTime="2025-10-02 09:46:50.612784081 +0000 UTC m=+1175.969128106" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.636025 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.656470 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.690423 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.691988 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.694871 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.695011 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2l9tg" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.695154 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.695303 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.702633 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.715506 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.727414 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.739016 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.750724 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.764087 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.775556 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.778741 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.781089 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.781385 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.802126 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.805695 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.807669 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.808245 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.851967 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.868710 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.887926 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a75ea0f7-0f30-4aa7-8608-78a743ed275b-logs\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.888115 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.888333 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.888438 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a75ea0f7-0f30-4aa7-8608-78a743ed275b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.888500 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnjzg\" (UniqueName: \"kubernetes.io/projected/a75ea0f7-0f30-4aa7-8608-78a743ed275b-kube-api-access-gnjzg\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.888578 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.888617 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.888736 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.989977 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-config-data\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990778 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990814 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990839 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-log-httpd\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990857 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-scripts\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990881 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990899 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990926 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990946 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a75ea0f7-0f30-4aa7-8608-78a743ed275b-logs\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.990980 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj6fq\" (UniqueName: \"kubernetes.io/projected/b4709e97-762f-439e-afe8-6617a9668863-kube-api-access-mj6fq\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991002 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991028 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-run-httpd\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991046 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-scripts\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991085 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991109 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/046bf1a9-09e7-4ea0-a26e-d6de90838c68-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991131 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2rcm\" (UniqueName: \"kubernetes.io/projected/046bf1a9-09e7-4ea0-a26e-d6de90838c68-kube-api-access-d2rcm\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991150 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991171 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-config-data\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991209 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991243 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a75ea0f7-0f30-4aa7-8608-78a743ed275b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991268 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046bf1a9-09e7-4ea0-a26e-d6de90838c68-logs\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991291 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.991310 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnjzg\" (UniqueName: \"kubernetes.io/projected/a75ea0f7-0f30-4aa7-8608-78a743ed275b-kube-api-access-gnjzg\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.992298 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.993306 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a75ea0f7-0f30-4aa7-8608-78a743ed275b-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:50 crc kubenswrapper[5035]: I1002 09:46:50.993665 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a75ea0f7-0f30-4aa7-8608-78a743ed275b-logs\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:50.998275 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.008588 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.008742 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.008900 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75ea0f7-0f30-4aa7-8608-78a743ed275b-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.019679 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnjzg\" (UniqueName: \"kubernetes.io/projected/a75ea0f7-0f30-4aa7-8608-78a743ed275b-kube-api-access-gnjzg\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.044263 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"a75ea0f7-0f30-4aa7-8608-78a743ed275b\") " pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093362 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj6fq\" (UniqueName: \"kubernetes.io/projected/b4709e97-762f-439e-afe8-6617a9668863-kube-api-access-mj6fq\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093415 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-run-httpd\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093472 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-scripts\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093521 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093577 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/046bf1a9-09e7-4ea0-a26e-d6de90838c68-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093603 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2rcm\" (UniqueName: \"kubernetes.io/projected/046bf1a9-09e7-4ea0-a26e-d6de90838c68-kube-api-access-d2rcm\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093633 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-config-data\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093703 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046bf1a9-09e7-4ea0-a26e-d6de90838c68-logs\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093739 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093763 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-config-data\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093804 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-log-httpd\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093827 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-scripts\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093858 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.093877 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.094291 5035 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.095120 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/046bf1a9-09e7-4ea0-a26e-d6de90838c68-logs\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.095787 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-log-httpd\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.096056 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/046bf1a9-09e7-4ea0-a26e-d6de90838c68-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.096672 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-run-httpd\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.106122 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.107162 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-config-data\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.109346 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-scripts\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.109672 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-config-data\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.110074 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.110486 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.111035 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.111054 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/046bf1a9-09e7-4ea0-a26e-d6de90838c68-scripts\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.111783 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2rcm\" (UniqueName: \"kubernetes.io/projected/046bf1a9-09e7-4ea0-a26e-d6de90838c68-kube-api-access-d2rcm\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.113057 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj6fq\" (UniqueName: \"kubernetes.io/projected/b4709e97-762f-439e-afe8-6617a9668863-kube-api-access-mj6fq\") pod \"ceilometer-0\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.137047 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"046bf1a9-09e7-4ea0-a26e-d6de90838c68\") " pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.183707 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.196727 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.226573 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.601069 5035 generic.go:334] "Generic (PLEG): container finished" podID="dd4355e8-6dd5-4937-969a-bad47330d25b" containerID="19d62c924f4defa5c73f856ec6045a18c0f4e681a4f305c42a79d770f81a2f01" exitCode=0 Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.601860 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerDied","Data":"19d62c924f4defa5c73f856ec6045a18c0f4e681a4f305c42a79d770f81a2f01"} Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.601931 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerStarted","Data":"f2f5103d26800147adc325902830070f7a9d81b75d0e1763b3f91f558da4f140"} Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.878740 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 09:46:51 crc kubenswrapper[5035]: I1002 09:46:51.947464 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.079438 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.190246 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96e76475-0fec-46d0-8e6c-d346bfb3b456" path="/var/lib/kubelet/pods/96e76475-0fec-46d0-8e6c-d346bfb3b456/volumes" Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.192025 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc55ee7d-a2e9-4bc7-a86d-9d4d58356442" path="/var/lib/kubelet/pods/cc55ee7d-a2e9-4bc7-a86d-9d4d58356442/volumes" Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.193460 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f42b1109-c1b4-4a08-a8d3-03cc50774a90" path="/var/lib/kubelet/pods/f42b1109-c1b4-4a08-a8d3-03cc50774a90/volumes" Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.613503 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerStarted","Data":"d671ee8ccfb24b4d38a11ea6a6a81b8cab427cfaef4cc23b7fd2df9c32dc4fa8"} Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.619187 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a75ea0f7-0f30-4aa7-8608-78a743ed275b","Type":"ContainerStarted","Data":"ea6ce25848e8e7daaf165f749a317a979dec169c43226010f77448db00ddc32c"} Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.620251 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"046bf1a9-09e7-4ea0-a26e-d6de90838c68","Type":"ContainerStarted","Data":"0a8200763b4be8ee51aaca3e3bcbc8aeb17d5be74ba6f7675e31924f8e202391"} Oct 02 09:46:52 crc kubenswrapper[5035]: I1002 09:46:52.621132 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerStarted","Data":"3df54b51646f797ff32ca150ed8ec435681753bfcb496cafe2b008aa2717ba53"} Oct 02 09:46:53 crc kubenswrapper[5035]: I1002 09:46:53.631243 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerStarted","Data":"4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f"} Oct 02 09:46:53 crc kubenswrapper[5035]: I1002 09:46:53.637340 5035 generic.go:334] "Generic (PLEG): container finished" podID="dd4355e8-6dd5-4937-969a-bad47330d25b" containerID="d671ee8ccfb24b4d38a11ea6a6a81b8cab427cfaef4cc23b7fd2df9c32dc4fa8" exitCode=0 Oct 02 09:46:53 crc kubenswrapper[5035]: I1002 09:46:53.637421 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerDied","Data":"d671ee8ccfb24b4d38a11ea6a6a81b8cab427cfaef4cc23b7fd2df9c32dc4fa8"} Oct 02 09:46:53 crc kubenswrapper[5035]: I1002 09:46:53.644700 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a75ea0f7-0f30-4aa7-8608-78a743ed275b","Type":"ContainerStarted","Data":"0bafd640bcba71e89907e51db74824edafd25b6fb3db340cc851a06334781be8"} Oct 02 09:46:53 crc kubenswrapper[5035]: I1002 09:46:53.646919 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"046bf1a9-09e7-4ea0-a26e-d6de90838c68","Type":"ContainerStarted","Data":"244b2d119dee08eea7bb957492a8b3843b2d8a3c11ae1c71dd6ac7b1f05be003"} Oct 02 09:46:54 crc kubenswrapper[5035]: I1002 09:46:54.659740 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a75ea0f7-0f30-4aa7-8608-78a743ed275b","Type":"ContainerStarted","Data":"bbad785c6fadb17999e3f5695fc3d34ae2f9f5d1ec042d399185ba87fbb5788d"} Oct 02 09:46:54 crc kubenswrapper[5035]: I1002 09:46:54.661713 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"046bf1a9-09e7-4ea0-a26e-d6de90838c68","Type":"ContainerStarted","Data":"d7ec9432542530629adfc1d8acd3ee58db04148b549e851ddc29292b333d64e5"} Oct 02 09:46:54 crc kubenswrapper[5035]: I1002 09:46:54.664883 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerStarted","Data":"5791ea9c9103dad8d824dd654220022df07d185ef3947eba6008d561f86b4f39"} Oct 02 09:46:54 crc kubenswrapper[5035]: I1002 09:46:54.697263 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.697231789 podStartE2EDuration="4.697231789s" podCreationTimestamp="2025-10-02 09:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:46:54.687391381 +0000 UTC m=+1180.043735436" watchObservedRunningTime="2025-10-02 09:46:54.697231789 +0000 UTC m=+1180.053575834" Oct 02 09:46:54 crc kubenswrapper[5035]: I1002 09:46:54.726848 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.726809173 podStartE2EDuration="4.726809173s" podCreationTimestamp="2025-10-02 09:46:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:46:54.716931254 +0000 UTC m=+1180.073275319" watchObservedRunningTime="2025-10-02 09:46:54.726809173 +0000 UTC m=+1180.083153218" Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.538152 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.538576 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.538616 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.539082 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db9ffa0950502b83608fcd73436cfa37aeaa8a721d02dc03e8e35488a5442a33"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.539143 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://db9ffa0950502b83608fcd73436cfa37aeaa8a721d02dc03e8e35488a5442a33" gracePeriod=600 Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.690677 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerStarted","Data":"33b82af8ada5b4defb9b0c69035f06e653b50e58840bf9d265aa5b456f4b81a5"} Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.694270 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="db9ffa0950502b83608fcd73436cfa37aeaa8a721d02dc03e8e35488a5442a33" exitCode=0 Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.694348 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"db9ffa0950502b83608fcd73436cfa37aeaa8a721d02dc03e8e35488a5442a33"} Oct 02 09:46:55 crc kubenswrapper[5035]: I1002 09:46:55.694396 5035 scope.go:117] "RemoveContainer" containerID="a1bca98fb1c6a4355a4320624f5591fb78dba746011bbeaf90858ccf387d2ab3" Oct 02 09:46:56 crc kubenswrapper[5035]: I1002 09:46:56.705429 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"e6c5f2097a0a83c78bab17bcdbccb738feed82e531c37ccc7b2955b0ca92c739"} Oct 02 09:46:56 crc kubenswrapper[5035]: I1002 09:46:56.712272 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerStarted","Data":"bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416"} Oct 02 09:46:56 crc kubenswrapper[5035]: I1002 09:46:56.722487 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerStarted","Data":"9b90c37ab5f803760343b5075f386e5a5bf6317ad0fcdf780d06c167a7658d5c"} Oct 02 09:46:57 crc kubenswrapper[5035]: I1002 09:46:57.735632 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerStarted","Data":"3908934f2ad82f856664ed8106e7e22494a8880477c133c66238495bb453f1af"} Oct 02 09:46:57 crc kubenswrapper[5035]: I1002 09:46:57.735989 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 02 09:46:57 crc kubenswrapper[5035]: I1002 09:46:57.764481 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-0" podStartSLOduration=8.764460628 podStartE2EDuration="8.764460628s" podCreationTimestamp="2025-10-02 09:46:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:46:57.76351848 +0000 UTC m=+1183.119862505" watchObservedRunningTime="2025-10-02 09:46:57.764460628 +0000 UTC m=+1183.120804653" Oct 02 09:46:58 crc kubenswrapper[5035]: I1002 09:46:58.747405 5035 generic.go:334] "Generic (PLEG): container finished" podID="dd4355e8-6dd5-4937-969a-bad47330d25b" containerID="33b82af8ada5b4defb9b0c69035f06e653b50e58840bf9d265aa5b456f4b81a5" exitCode=0 Oct 02 09:46:58 crc kubenswrapper[5035]: I1002 09:46:58.747579 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerDied","Data":"33b82af8ada5b4defb9b0c69035f06e653b50e58840bf9d265aa5b456f4b81a5"} Oct 02 09:46:58 crc kubenswrapper[5035]: I1002 09:46:58.748328 5035 scope.go:117] "RemoveContainer" containerID="33b82af8ada5b4defb9b0c69035f06e653b50e58840bf9d265aa5b456f4b81a5" Oct 02 09:46:59 crc kubenswrapper[5035]: I1002 09:46:59.760654 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"dd4355e8-6dd5-4937-969a-bad47330d25b","Type":"ContainerStarted","Data":"7f7c0e03fe5b5abe44bcec8092aa1cb15f7f67cb4847fd37b44d1c18cdde3592"} Oct 02 09:46:59 crc kubenswrapper[5035]: I1002 09:46:59.762755 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerStarted","Data":"922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83"} Oct 02 09:47:00 crc kubenswrapper[5035]: I1002 09:47:00.182989 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Oct 02 09:47:00 crc kubenswrapper[5035]: I1002 09:47:00.183046 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Oct 02 09:47:00 crc kubenswrapper[5035]: I1002 09:47:00.183064 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 02 09:47:00 crc kubenswrapper[5035]: I1002 09:47:00.183079 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 02 09:47:00 crc kubenswrapper[5035]: I1002 09:47:00.189885 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="dd4355e8-6dd5-4937-969a-bad47330d25b" containerName="ironic-inspector" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 09:47:00 crc kubenswrapper[5035]: I1002 09:47:00.196910 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="dd4355e8-6dd5-4937-969a-bad47330d25b" containerName="ironic-inspector-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.184404 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.185653 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.223203 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.223688 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.227045 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.227122 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.273736 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.280211 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.790126 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerStarted","Data":"53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e"} Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.791335 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.791380 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.791397 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.791412 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.792678 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 02 09:47:01 crc kubenswrapper[5035]: I1002 09:47:01.824715 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.285348493 podStartE2EDuration="11.824689788s" podCreationTimestamp="2025-10-02 09:46:50 +0000 UTC" firstStartedPulling="2025-10-02 09:46:52.008788383 +0000 UTC m=+1177.365132408" lastFinishedPulling="2025-10-02 09:47:00.548129678 +0000 UTC m=+1185.904473703" observedRunningTime="2025-10-02 09:47:01.810920966 +0000 UTC m=+1187.167264991" watchObservedRunningTime="2025-10-02 09:47:01.824689788 +0000 UTC m=+1187.181033813" Oct 02 09:47:02 crc kubenswrapper[5035]: I1002 09:47:02.799677 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:47:03 crc kubenswrapper[5035]: I1002 09:47:03.811001 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:47:03 crc kubenswrapper[5035]: I1002 09:47:03.811360 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:47:04 crc kubenswrapper[5035]: I1002 09:47:04.021928 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:04 crc kubenswrapper[5035]: I1002 09:47:04.021991 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 09:47:04 crc kubenswrapper[5035]: I1002 09:47:04.022049 5035 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 09:47:04 crc kubenswrapper[5035]: I1002 09:47:04.033054 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 09:47:04 crc kubenswrapper[5035]: I1002 09:47:04.074327 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 09:47:10 crc kubenswrapper[5035]: I1002 09:47:10.210093 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Oct 02 09:47:10 crc kubenswrapper[5035]: I1002 09:47:10.212384 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Oct 02 09:47:10 crc kubenswrapper[5035]: I1002 09:47:10.217892 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 02 09:47:10 crc kubenswrapper[5035]: I1002 09:47:10.902598 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 02 09:47:21 crc kubenswrapper[5035]: I1002 09:47:21.201653 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 09:47:21 crc kubenswrapper[5035]: I1002 09:47:21.945927 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:21 crc kubenswrapper[5035]: I1002 09:47:21.994247 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-central-agent" containerID="cri-o://4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f" gracePeriod=30 Oct 02 09:47:21 crc kubenswrapper[5035]: I1002 09:47:21.994362 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="proxy-httpd" containerID="cri-o://53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e" gracePeriod=30 Oct 02 09:47:21 crc kubenswrapper[5035]: I1002 09:47:21.994381 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-notification-agent" containerID="cri-o://bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416" gracePeriod=30 Oct 02 09:47:21 crc kubenswrapper[5035]: I1002 09:47:21.994362 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="sg-core" containerID="cri-o://922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83" gracePeriod=30 Oct 02 09:47:23 crc kubenswrapper[5035]: I1002 09:47:23.006473 5035 generic.go:334] "Generic (PLEG): container finished" podID="b4709e97-762f-439e-afe8-6617a9668863" containerID="53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e" exitCode=0 Oct 02 09:47:23 crc kubenswrapper[5035]: I1002 09:47:23.006775 5035 generic.go:334] "Generic (PLEG): container finished" podID="b4709e97-762f-439e-afe8-6617a9668863" containerID="922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83" exitCode=2 Oct 02 09:47:23 crc kubenswrapper[5035]: I1002 09:47:23.006784 5035 generic.go:334] "Generic (PLEG): container finished" podID="b4709e97-762f-439e-afe8-6617a9668863" containerID="4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f" exitCode=0 Oct 02 09:47:23 crc kubenswrapper[5035]: I1002 09:47:23.006570 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerDied","Data":"53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e"} Oct 02 09:47:23 crc kubenswrapper[5035]: I1002 09:47:23.006821 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerDied","Data":"922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83"} Oct 02 09:47:23 crc kubenswrapper[5035]: I1002 09:47:23.006835 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerDied","Data":"4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f"} Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.220334 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.220921 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f677fa29-64c1-45f7-8f6d-c4c15f83db60" containerName="kube-state-metrics" containerID="cri-o://4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535" gracePeriod=30 Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.769643 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.828445 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.921670 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj6fq\" (UniqueName: \"kubernetes.io/projected/b4709e97-762f-439e-afe8-6617a9668863-kube-api-access-mj6fq\") pod \"b4709e97-762f-439e-afe8-6617a9668863\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.921756 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-sg-core-conf-yaml\") pod \"b4709e97-762f-439e-afe8-6617a9668863\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.921820 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-run-httpd\") pod \"b4709e97-762f-439e-afe8-6617a9668863\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.921857 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-combined-ca-bundle\") pod \"b4709e97-762f-439e-afe8-6617a9668863\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.921927 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-log-httpd\") pod \"b4709e97-762f-439e-afe8-6617a9668863\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.921995 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-scripts\") pod \"b4709e97-762f-439e-afe8-6617a9668863\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.922034 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-config-data\") pod \"b4709e97-762f-439e-afe8-6617a9668863\" (UID: \"b4709e97-762f-439e-afe8-6617a9668863\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.922055 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkfmf\" (UniqueName: \"kubernetes.io/projected/f677fa29-64c1-45f7-8f6d-c4c15f83db60-kube-api-access-kkfmf\") pod \"f677fa29-64c1-45f7-8f6d-c4c15f83db60\" (UID: \"f677fa29-64c1-45f7-8f6d-c4c15f83db60\") " Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.922185 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b4709e97-762f-439e-afe8-6617a9668863" (UID: "b4709e97-762f-439e-afe8-6617a9668863"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.922408 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.923115 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b4709e97-762f-439e-afe8-6617a9668863" (UID: "b4709e97-762f-439e-afe8-6617a9668863"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.927851 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4709e97-762f-439e-afe8-6617a9668863-kube-api-access-mj6fq" (OuterVolumeSpecName: "kube-api-access-mj6fq") pod "b4709e97-762f-439e-afe8-6617a9668863" (UID: "b4709e97-762f-439e-afe8-6617a9668863"). InnerVolumeSpecName "kube-api-access-mj6fq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.928234 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f677fa29-64c1-45f7-8f6d-c4c15f83db60-kube-api-access-kkfmf" (OuterVolumeSpecName: "kube-api-access-kkfmf") pod "f677fa29-64c1-45f7-8f6d-c4c15f83db60" (UID: "f677fa29-64c1-45f7-8f6d-c4c15f83db60"). InnerVolumeSpecName "kube-api-access-kkfmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.928301 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-scripts" (OuterVolumeSpecName: "scripts") pod "b4709e97-762f-439e-afe8-6617a9668863" (UID: "b4709e97-762f-439e-afe8-6617a9668863"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:26 crc kubenswrapper[5035]: I1002 09:47:26.953933 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b4709e97-762f-439e-afe8-6617a9668863" (UID: "b4709e97-762f-439e-afe8-6617a9668863"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.014345 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4709e97-762f-439e-afe8-6617a9668863" (UID: "b4709e97-762f-439e-afe8-6617a9668863"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.024679 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.025056 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkfmf\" (UniqueName: \"kubernetes.io/projected/f677fa29-64c1-45f7-8f6d-c4c15f83db60-kube-api-access-kkfmf\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.025155 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj6fq\" (UniqueName: \"kubernetes.io/projected/b4709e97-762f-439e-afe8-6617a9668863-kube-api-access-mj6fq\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.025245 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.025329 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.025412 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b4709e97-762f-439e-afe8-6617a9668863-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.029470 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-config-data" (OuterVolumeSpecName: "config-data") pod "b4709e97-762f-439e-afe8-6617a9668863" (UID: "b4709e97-762f-439e-afe8-6617a9668863"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.048867 5035 generic.go:334] "Generic (PLEG): container finished" podID="b4709e97-762f-439e-afe8-6617a9668863" containerID="bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416" exitCode=0 Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.048934 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.048959 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerDied","Data":"bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416"} Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.049320 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b4709e97-762f-439e-afe8-6617a9668863","Type":"ContainerDied","Data":"3df54b51646f797ff32ca150ed8ec435681753bfcb496cafe2b008aa2717ba53"} Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.049356 5035 scope.go:117] "RemoveContainer" containerID="53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.051991 5035 generic.go:334] "Generic (PLEG): container finished" podID="f677fa29-64c1-45f7-8f6d-c4c15f83db60" containerID="4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535" exitCode=2 Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.052034 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f677fa29-64c1-45f7-8f6d-c4c15f83db60","Type":"ContainerDied","Data":"4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535"} Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.052059 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f677fa29-64c1-45f7-8f6d-c4c15f83db60","Type":"ContainerDied","Data":"ffa09eef49ab5dff9dcf76901125979935e5a30ff6757fc7afc1a38c4a3eb629"} Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.052105 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.084116 5035 scope.go:117] "RemoveContainer" containerID="922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.106391 5035 scope.go:117] "RemoveContainer" containerID="bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.107822 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.120548 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.126865 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4709e97-762f-439e-afe8-6617a9668863-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.144111 5035 scope.go:117] "RemoveContainer" containerID="4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.148657 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.149095 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="sg-core" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149109 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="sg-core" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.149130 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="proxy-httpd" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149136 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="proxy-httpd" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.149149 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-notification-agent" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149158 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-notification-agent" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.149171 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f677fa29-64c1-45f7-8f6d-c4c15f83db60" containerName="kube-state-metrics" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149177 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f677fa29-64c1-45f7-8f6d-c4c15f83db60" containerName="kube-state-metrics" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.149190 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-central-agent" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149195 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-central-agent" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149412 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f677fa29-64c1-45f7-8f6d-c4c15f83db60" containerName="kube-state-metrics" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149431 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="sg-core" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149442 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-central-agent" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149455 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="proxy-httpd" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.149471 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4709e97-762f-439e-afe8-6617a9668863" containerName="ceilometer-notification-agent" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.150146 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.154546 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.154623 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.154987 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-btqtx" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.159468 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.174251 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.182127 5035 scope.go:117] "RemoveContainer" containerID="53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.182294 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e\": container with ID starting with 53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e not found: ID does not exist" containerID="53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.182326 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e"} err="failed to get container status \"53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e\": rpc error: code = NotFound desc = could not find container \"53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e\": container with ID starting with 53961cc7b7fb0c22e6c8ae2fa8e224bc8f774d3c4fd796e7a3ddf79e7995fc4e not found: ID does not exist" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.182355 5035 scope.go:117] "RemoveContainer" containerID="922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.182499 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83\": container with ID starting with 922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83 not found: ID does not exist" containerID="922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.182513 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83"} err="failed to get container status \"922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83\": rpc error: code = NotFound desc = could not find container \"922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83\": container with ID starting with 922300e9f0b2c1c316719396bbb8d31b9b5937560c069397c7ed18a61b401e83 not found: ID does not exist" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.182524 5035 scope.go:117] "RemoveContainer" containerID="bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.182662 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416\": container with ID starting with bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416 not found: ID does not exist" containerID="bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.182688 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416"} err="failed to get container status \"bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416\": rpc error: code = NotFound desc = could not find container \"bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416\": container with ID starting with bfe7723769859f7b6042b8b4d6b190354f2798b69911335f0c40e9014b67e416 not found: ID does not exist" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.182699 5035 scope.go:117] "RemoveContainer" containerID="4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f" Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.182971 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f\": container with ID starting with 4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f not found: ID does not exist" containerID="4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.183022 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f"} err="failed to get container status \"4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f\": rpc error: code = NotFound desc = could not find container \"4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f\": container with ID starting with 4d291d84b2261749fb7be93b42daa9949363be5c8e2a772457f563042d0e9b4f not found: ID does not exist" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.183080 5035 scope.go:117] "RemoveContainer" containerID="4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.192308 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.202087 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.212177 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.212371 5035 scope.go:117] "RemoveContainer" containerID="4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.213511 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: E1002 09:47:27.214714 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535\": container with ID starting with 4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535 not found: ID does not exist" containerID="4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.214748 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535"} err="failed to get container status \"4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535\": rpc error: code = NotFound desc = could not find container \"4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535\": container with ID starting with 4b1e806e678d3b39c799ab37a1c683da8bb368c683e10e3160f0679104727535 not found: ID does not exist" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.215368 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.215493 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.228208 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.228271 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz8vn\" (UniqueName: \"kubernetes.io/projected/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-api-access-qz8vn\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.228307 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.228342 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330204 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-log-httpd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330348 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330450 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330495 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz8vn\" (UniqueName: \"kubernetes.io/projected/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-api-access-qz8vn\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330513 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330558 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-scripts\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330575 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-run-httpd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330593 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330611 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-config-data\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330650 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.330669 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6qwd\" (UniqueName: \"kubernetes.io/projected/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-kube-api-access-w6qwd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.335245 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.335334 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.335757 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.347318 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz8vn\" (UniqueName: \"kubernetes.io/projected/e78db9f0-3ff7-444c-9cea-510d690eafc5-kube-api-access-qz8vn\") pod \"kube-state-metrics-0\" (UID: \"e78db9f0-3ff7-444c-9cea-510d690eafc5\") " pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.432281 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.432331 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-scripts\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.432357 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-run-httpd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.432385 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-config-data\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.432421 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6qwd\" (UniqueName: \"kubernetes.io/projected/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-kube-api-access-w6qwd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.432477 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-log-httpd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.432575 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.433238 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-run-httpd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.433817 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-log-httpd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.437994 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-config-data\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.438065 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-scripts\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.438197 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.438488 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.450287 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6qwd\" (UniqueName: \"kubernetes.io/projected/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-kube-api-access-w6qwd\") pod \"ceilometer-0\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.492283 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.589141 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:27 crc kubenswrapper[5035]: I1002 09:47:27.950139 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 09:47:27 crc kubenswrapper[5035]: W1002 09:47:27.953102 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode78db9f0_3ff7_444c_9cea_510d690eafc5.slice/crio-086cc401214c4e248f1b341b938c16600a3500e10cf560d77e9635d30709a103 WatchSource:0}: Error finding container 086cc401214c4e248f1b341b938c16600a3500e10cf560d77e9635d30709a103: Status 404 returned error can't find the container with id 086cc401214c4e248f1b341b938c16600a3500e10cf560d77e9635d30709a103 Oct 02 09:47:28 crc kubenswrapper[5035]: I1002 09:47:28.057806 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:28 crc kubenswrapper[5035]: I1002 09:47:28.069921 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e78db9f0-3ff7-444c-9cea-510d690eafc5","Type":"ContainerStarted","Data":"086cc401214c4e248f1b341b938c16600a3500e10cf560d77e9635d30709a103"} Oct 02 09:47:28 crc kubenswrapper[5035]: I1002 09:47:28.175246 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4709e97-762f-439e-afe8-6617a9668863" path="/var/lib/kubelet/pods/b4709e97-762f-439e-afe8-6617a9668863/volumes" Oct 02 09:47:28 crc kubenswrapper[5035]: I1002 09:47:28.176300 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f677fa29-64c1-45f7-8f6d-c4c15f83db60" path="/var/lib/kubelet/pods/f677fa29-64c1-45f7-8f6d-c4c15f83db60/volumes" Oct 02 09:47:28 crc kubenswrapper[5035]: I1002 09:47:28.538209 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:29 crc kubenswrapper[5035]: I1002 09:47:29.083204 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerStarted","Data":"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269"} Oct 02 09:47:29 crc kubenswrapper[5035]: I1002 09:47:29.083543 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerStarted","Data":"866749bd2adee9b52ad01b22aeb512312b2548eb8de3205702278afd46f59148"} Oct 02 09:47:29 crc kubenswrapper[5035]: I1002 09:47:29.084870 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e78db9f0-3ff7-444c-9cea-510d690eafc5","Type":"ContainerStarted","Data":"c8b34fd1e5d2cec2a3d488f7ddee19885fcae4a667e485b01713a76777db124e"} Oct 02 09:47:29 crc kubenswrapper[5035]: I1002 09:47:29.084995 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 09:47:29 crc kubenswrapper[5035]: I1002 09:47:29.108058 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.2861080999999999 podStartE2EDuration="2.108035428s" podCreationTimestamp="2025-10-02 09:47:27 +0000 UTC" firstStartedPulling="2025-10-02 09:47:27.955113193 +0000 UTC m=+1213.311457218" lastFinishedPulling="2025-10-02 09:47:28.777040521 +0000 UTC m=+1214.133384546" observedRunningTime="2025-10-02 09:47:29.099334303 +0000 UTC m=+1214.455678328" watchObservedRunningTime="2025-10-02 09:47:29.108035428 +0000 UTC m=+1214.464379453" Oct 02 09:47:30 crc kubenswrapper[5035]: I1002 09:47:30.132806 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerStarted","Data":"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659"} Oct 02 09:47:31 crc kubenswrapper[5035]: I1002 09:47:31.143792 5035 generic.go:334] "Generic (PLEG): container finished" podID="54ae7cc3-99fc-4485-b5b3-589d64791997" containerID="7f617c73dfbdcae4fe1152dd876e3ad4da1920b515ac23133acb5a9b3270619c" exitCode=0 Oct 02 09:47:31 crc kubenswrapper[5035]: I1002 09:47:31.143834 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" event={"ID":"54ae7cc3-99fc-4485-b5b3-589d64791997","Type":"ContainerDied","Data":"7f617c73dfbdcae4fe1152dd876e3ad4da1920b515ac23133acb5a9b3270619c"} Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.155710 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerStarted","Data":"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47"} Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.510170 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.644519 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvk5m\" (UniqueName: \"kubernetes.io/projected/54ae7cc3-99fc-4485-b5b3-589d64791997-kube-api-access-nvk5m\") pod \"54ae7cc3-99fc-4485-b5b3-589d64791997\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.644891 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-config-data\") pod \"54ae7cc3-99fc-4485-b5b3-589d64791997\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.644933 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-combined-ca-bundle\") pod \"54ae7cc3-99fc-4485-b5b3-589d64791997\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.644976 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-scripts\") pod \"54ae7cc3-99fc-4485-b5b3-589d64791997\" (UID: \"54ae7cc3-99fc-4485-b5b3-589d64791997\") " Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.650749 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ae7cc3-99fc-4485-b5b3-589d64791997-kube-api-access-nvk5m" (OuterVolumeSpecName: "kube-api-access-nvk5m") pod "54ae7cc3-99fc-4485-b5b3-589d64791997" (UID: "54ae7cc3-99fc-4485-b5b3-589d64791997"). InnerVolumeSpecName "kube-api-access-nvk5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.661274 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-scripts" (OuterVolumeSpecName: "scripts") pod "54ae7cc3-99fc-4485-b5b3-589d64791997" (UID: "54ae7cc3-99fc-4485-b5b3-589d64791997"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.671970 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54ae7cc3-99fc-4485-b5b3-589d64791997" (UID: "54ae7cc3-99fc-4485-b5b3-589d64791997"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.690333 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-config-data" (OuterVolumeSpecName: "config-data") pod "54ae7cc3-99fc-4485-b5b3-589d64791997" (UID: "54ae7cc3-99fc-4485-b5b3-589d64791997"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.747740 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvk5m\" (UniqueName: \"kubernetes.io/projected/54ae7cc3-99fc-4485-b5b3-589d64791997-kube-api-access-nvk5m\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.747784 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.747799 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:32 crc kubenswrapper[5035]: I1002 09:47:32.747809 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54ae7cc3-99fc-4485-b5b3-589d64791997-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.166004 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" event={"ID":"54ae7cc3-99fc-4485-b5b3-589d64791997","Type":"ContainerDied","Data":"7ca3b7eb4081062e497241b3ceeecf35c99e45101ab0e95e4dc3d2ad70863db1"} Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.166034 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ca3b7eb4081062e497241b3ceeecf35c99e45101ab0e95e4dc3d2ad70863db1" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.166079 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sgtqw" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.249757 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 09:47:33 crc kubenswrapper[5035]: E1002 09:47:33.250154 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ae7cc3-99fc-4485-b5b3-589d64791997" containerName="nova-cell0-conductor-db-sync" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.250172 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ae7cc3-99fc-4485-b5b3-589d64791997" containerName="nova-cell0-conductor-db-sync" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.250358 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ae7cc3-99fc-4485-b5b3-589d64791997" containerName="nova-cell0-conductor-db-sync" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.251028 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.254852 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.255251 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-b62ln" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.259763 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.360057 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwsc7\" (UniqueName: \"kubernetes.io/projected/03343b82-9ffd-4ab8-9b8e-88a1277d3014-kube-api-access-gwsc7\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.360184 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03343b82-9ffd-4ab8-9b8e-88a1277d3014-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.360325 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03343b82-9ffd-4ab8-9b8e-88a1277d3014-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.462279 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwsc7\" (UniqueName: \"kubernetes.io/projected/03343b82-9ffd-4ab8-9b8e-88a1277d3014-kube-api-access-gwsc7\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.462352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03343b82-9ffd-4ab8-9b8e-88a1277d3014-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.462425 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03343b82-9ffd-4ab8-9b8e-88a1277d3014-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.466995 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03343b82-9ffd-4ab8-9b8e-88a1277d3014-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.471116 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03343b82-9ffd-4ab8-9b8e-88a1277d3014-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.485672 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwsc7\" (UniqueName: \"kubernetes.io/projected/03343b82-9ffd-4ab8-9b8e-88a1277d3014-kube-api-access-gwsc7\") pod \"nova-cell0-conductor-0\" (UID: \"03343b82-9ffd-4ab8-9b8e-88a1277d3014\") " pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:33 crc kubenswrapper[5035]: I1002 09:47:33.569525 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:34 crc kubenswrapper[5035]: I1002 09:47:34.020349 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 09:47:34 crc kubenswrapper[5035]: I1002 09:47:34.175316 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"03343b82-9ffd-4ab8-9b8e-88a1277d3014","Type":"ContainerStarted","Data":"8a2c88fd4bd50cdcfa8506b37401f5b74122113591e5385446597db0e1647f5c"} Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.188144 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"03343b82-9ffd-4ab8-9b8e-88a1277d3014","Type":"ContainerStarted","Data":"9b479b0e97a0bf97bede1f5eae9908a7efa38a95b2e174323855f5ffa7f5548d"} Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.190197 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.192196 5035 generic.go:334] "Generic (PLEG): container finished" podID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerID="5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28" exitCode=1 Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.192259 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerDied","Data":"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28"} Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.192429 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-central-agent" containerID="cri-o://8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269" gracePeriod=30 Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.192572 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="sg-core" containerID="cri-o://73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47" gracePeriod=30 Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.192633 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-notification-agent" containerID="cri-o://ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659" gracePeriod=30 Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.222864 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.222846173 podStartE2EDuration="2.222846173s" podCreationTimestamp="2025-10-02 09:47:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:47:35.214646844 +0000 UTC m=+1220.570990879" watchObservedRunningTime="2025-10-02 09:47:35.222846173 +0000 UTC m=+1220.579190198" Oct 02 09:47:35 crc kubenswrapper[5035]: I1002 09:47:35.971411 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.129491 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6qwd\" (UniqueName: \"kubernetes.io/projected/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-kube-api-access-w6qwd\") pod \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.129613 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-combined-ca-bundle\") pod \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.129762 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-run-httpd\") pod \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.129899 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-config-data\") pod \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.129990 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-sg-core-conf-yaml\") pod \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.130065 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-scripts\") pod \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.130089 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-log-httpd\") pod \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\" (UID: \"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a\") " Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.131144 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" (UID: "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.132389 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" (UID: "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.139217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-scripts" (OuterVolumeSpecName: "scripts") pod "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" (UID: "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.146120 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-kube-api-access-w6qwd" (OuterVolumeSpecName: "kube-api-access-w6qwd") pod "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" (UID: "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a"). InnerVolumeSpecName "kube-api-access-w6qwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.180743 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" (UID: "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.206054 5035 generic.go:334] "Generic (PLEG): container finished" podID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerID="73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47" exitCode=2 Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.206082 5035 generic.go:334] "Generic (PLEG): container finished" podID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerID="ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659" exitCode=0 Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.206090 5035 generic.go:334] "Generic (PLEG): container finished" podID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerID="8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269" exitCode=0 Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.207156 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.207695 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerDied","Data":"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47"} Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.207765 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerDied","Data":"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659"} Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.207779 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerDied","Data":"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269"} Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.207791 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a","Type":"ContainerDied","Data":"866749bd2adee9b52ad01b22aeb512312b2548eb8de3205702278afd46f59148"} Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.207811 5035 scope.go:117] "RemoveContainer" containerID="5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.232093 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.232408 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.232423 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.232433 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.232445 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6qwd\" (UniqueName: \"kubernetes.io/projected/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-kube-api-access-w6qwd\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.234440 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" (UID: "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.235637 5035 scope.go:117] "RemoveContainer" containerID="73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.255173 5035 scope.go:117] "RemoveContainer" containerID="ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.268526 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-config-data" (OuterVolumeSpecName: "config-data") pod "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" (UID: "c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.278724 5035 scope.go:117] "RemoveContainer" containerID="8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.303937 5035 scope.go:117] "RemoveContainer" containerID="5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28" Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.304453 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28\": container with ID starting with 5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28 not found: ID does not exist" containerID="5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.304506 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28"} err="failed to get container status \"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28\": rpc error: code = NotFound desc = could not find container \"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28\": container with ID starting with 5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.304528 5035 scope.go:117] "RemoveContainer" containerID="73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47" Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.304864 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47\": container with ID starting with 73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47 not found: ID does not exist" containerID="73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.304885 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47"} err="failed to get container status \"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47\": rpc error: code = NotFound desc = could not find container \"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47\": container with ID starting with 73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.304900 5035 scope.go:117] "RemoveContainer" containerID="ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659" Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.305200 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659\": container with ID starting with ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659 not found: ID does not exist" containerID="ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.305225 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659"} err="failed to get container status \"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659\": rpc error: code = NotFound desc = could not find container \"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659\": container with ID starting with ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.305239 5035 scope.go:117] "RemoveContainer" containerID="8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269" Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.305451 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269\": container with ID starting with 8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269 not found: ID does not exist" containerID="8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.305473 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269"} err="failed to get container status \"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269\": rpc error: code = NotFound desc = could not find container \"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269\": container with ID starting with 8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.305488 5035 scope.go:117] "RemoveContainer" containerID="5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.305689 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28"} err="failed to get container status \"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28\": rpc error: code = NotFound desc = could not find container \"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28\": container with ID starting with 5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.305707 5035 scope.go:117] "RemoveContainer" containerID="73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306008 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47"} err="failed to get container status \"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47\": rpc error: code = NotFound desc = could not find container \"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47\": container with ID starting with 73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306025 5035 scope.go:117] "RemoveContainer" containerID="ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306286 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659"} err="failed to get container status \"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659\": rpc error: code = NotFound desc = could not find container \"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659\": container with ID starting with ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306303 5035 scope.go:117] "RemoveContainer" containerID="8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306479 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269"} err="failed to get container status \"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269\": rpc error: code = NotFound desc = could not find container \"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269\": container with ID starting with 8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306497 5035 scope.go:117] "RemoveContainer" containerID="5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306839 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28"} err="failed to get container status \"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28\": rpc error: code = NotFound desc = could not find container \"5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28\": container with ID starting with 5378f1e1f9a8a0ffe8ae207a1570762712ebecc8a1875544f0f6e86a584eac28 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.306857 5035 scope.go:117] "RemoveContainer" containerID="73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.307071 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47"} err="failed to get container status \"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47\": rpc error: code = NotFound desc = could not find container \"73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47\": container with ID starting with 73faf8877e923cbfad429424e2544dd658f04998d19abac2984bd601e6e48c47 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.307088 5035 scope.go:117] "RemoveContainer" containerID="ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.307394 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659"} err="failed to get container status \"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659\": rpc error: code = NotFound desc = could not find container \"ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659\": container with ID starting with ca71b752963bb47660fc4561ba894f67b9a9625ffc2db4e007cde3d918c55659 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.307432 5035 scope.go:117] "RemoveContainer" containerID="8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.307630 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269"} err="failed to get container status \"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269\": rpc error: code = NotFound desc = could not find container \"8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269\": container with ID starting with 8ddc6b0a4788e2b455413665dace6030853b5a6f5fba2afc32442cddf5575269 not found: ID does not exist" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.333796 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.333824 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.600379 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.610182 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.624937 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.625595 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-notification-agent" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.625716 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-notification-agent" Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.625840 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-central-agent" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.625899 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-central-agent" Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.625959 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="proxy-httpd" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.626012 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="proxy-httpd" Oct 02 09:47:36 crc kubenswrapper[5035]: E1002 09:47:36.626078 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="sg-core" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.626137 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="sg-core" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.626438 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-notification-agent" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.626533 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="sg-core" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.626643 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="ceilometer-central-agent" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.626741 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" containerName="proxy-httpd" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.628956 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.631285 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.631523 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.641868 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.645755 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741135 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741218 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-config-data\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741410 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741469 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741499 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hblkg\" (UniqueName: \"kubernetes.io/projected/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-kube-api-access-hblkg\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741628 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-scripts\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741679 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.741754 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844108 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-scripts\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844198 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844267 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844348 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844435 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-config-data\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844497 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844522 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844583 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hblkg\" (UniqueName: \"kubernetes.io/projected/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-kube-api-access-hblkg\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844989 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-log-httpd\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.844999 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-run-httpd\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.860931 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.861007 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.861167 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-config-data\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.862310 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.862418 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-scripts\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.865327 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hblkg\" (UniqueName: \"kubernetes.io/projected/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-kube-api-access-hblkg\") pod \"ceilometer-0\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " pod="openstack/ceilometer-0" Oct 02 09:47:36 crc kubenswrapper[5035]: I1002 09:47:36.955264 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:47:37 crc kubenswrapper[5035]: I1002 09:47:37.427472 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:47:37 crc kubenswrapper[5035]: I1002 09:47:37.501571 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 09:47:38 crc kubenswrapper[5035]: I1002 09:47:38.173790 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a" path="/var/lib/kubelet/pods/c5bc104e-e2b6-4a36-a8f9-c7cbf9ca0d9a/volumes" Oct 02 09:47:38 crc kubenswrapper[5035]: I1002 09:47:38.239259 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerStarted","Data":"3887d0dfafbf976500675857b452965300326f20b7255bb676db9b6a329f62af"} Oct 02 09:47:39 crc kubenswrapper[5035]: I1002 09:47:39.253159 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerStarted","Data":"0768287e03f6c2b26503dc517804a61b5f0dac3818c4593be9246059e8f3a7fa"} Oct 02 09:47:39 crc kubenswrapper[5035]: I1002 09:47:39.253214 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerStarted","Data":"7a0d7223cb9ab44c4050171021800a688f29f1adc2a5d23300b979ae7da53649"} Oct 02 09:47:41 crc kubenswrapper[5035]: I1002 09:47:41.278057 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerStarted","Data":"25cfb55874b0f5f5155071b0d0f5babe50bb2c756a12cdaad26e799c30b34675"} Oct 02 09:47:42 crc kubenswrapper[5035]: I1002 09:47:42.287568 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerStarted","Data":"0f5c7f8199a7e7defdc74a072725c05bfe2e2b6fb5c72fbb2f025ad1d24fc416"} Oct 02 09:47:42 crc kubenswrapper[5035]: I1002 09:47:42.288871 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:47:42 crc kubenswrapper[5035]: I1002 09:47:42.326088 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.974522906 podStartE2EDuration="6.326068583s" podCreationTimestamp="2025-10-02 09:47:36 +0000 UTC" firstStartedPulling="2025-10-02 09:47:37.437131436 +0000 UTC m=+1222.793475461" lastFinishedPulling="2025-10-02 09:47:41.788677113 +0000 UTC m=+1227.145021138" observedRunningTime="2025-10-02 09:47:42.322245942 +0000 UTC m=+1227.678589977" watchObservedRunningTime="2025-10-02 09:47:42.326068583 +0000 UTC m=+1227.682412608" Oct 02 09:47:43 crc kubenswrapper[5035]: I1002 09:47:43.600146 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.056138 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-7jwqw"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.057798 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.061134 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.061277 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.074812 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7jwqw"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.184985 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.185387 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-scripts\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.185497 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9wps\" (UniqueName: \"kubernetes.io/projected/104b87a6-20cd-459e-b095-eaa7de00ba31-kube-api-access-h9wps\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.185716 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-config-data\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.199006 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.201202 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.203973 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.217422 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.280453 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.281861 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.288659 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.288938 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-config-data\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.289082 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.289231 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2d2t\" (UniqueName: \"kubernetes.io/projected/6ac1551c-b8d0-4141-b334-2bcdc550cce8-kube-api-access-c2d2t\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.289305 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac1551c-b8d0-4141-b334-2bcdc550cce8-logs\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.289409 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-scripts\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.289499 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9wps\" (UniqueName: \"kubernetes.io/projected/104b87a6-20cd-459e-b095-eaa7de00ba31-kube-api-access-h9wps\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.289661 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-config-data\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.292028 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.294840 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.303857 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-scripts\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.327803 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.328503 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-config-data\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.330897 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9wps\" (UniqueName: \"kubernetes.io/projected/104b87a6-20cd-459e-b095-eaa7de00ba31-kube-api-access-h9wps\") pod \"nova-cell0-cell-mapping-7jwqw\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.343613 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.345809 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.350337 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.386503 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.395195 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.395352 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-config-data\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.395438 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnntj\" (UniqueName: \"kubernetes.io/projected/00e17dd1-f431-4b62-aaac-0f6d466870e4-kube-api-access-cnntj\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.395578 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.395698 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-config-data\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.395793 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2d2t\" (UniqueName: \"kubernetes.io/projected/6ac1551c-b8d0-4141-b334-2bcdc550cce8-kube-api-access-c2d2t\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.395868 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac1551c-b8d0-4141-b334-2bcdc550cce8-logs\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.396262 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac1551c-b8d0-4141-b334-2bcdc550cce8-logs\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.429311 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2d2t\" (UniqueName: \"kubernetes.io/projected/6ac1551c-b8d0-4141-b334-2bcdc550cce8-kube-api-access-c2d2t\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.429339 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-config-data\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.432423 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.467089 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.497444 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.497502 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnntj\" (UniqueName: \"kubernetes.io/projected/00e17dd1-f431-4b62-aaac-0f6d466870e4-kube-api-access-cnntj\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.497590 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176d4726-3a2c-4e29-b6d8-1dc7139013a1-logs\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.497613 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkbzb\" (UniqueName: \"kubernetes.io/projected/176d4726-3a2c-4e29-b6d8-1dc7139013a1-kube-api-access-mkbzb\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.497654 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-config-data\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.497689 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-config-data\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.497768 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.514554 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.519292 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-config-data\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.530030 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.539751 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-759js"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.560791 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnntj\" (UniqueName: \"kubernetes.io/projected/00e17dd1-f431-4b62-aaac-0f6d466870e4-kube-api-access-cnntj\") pod \"nova-scheduler-0\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.570114 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.596250 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-759js"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.607324 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.613571 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176d4726-3a2c-4e29-b6d8-1dc7139013a1-logs\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.613674 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkbzb\" (UniqueName: \"kubernetes.io/projected/176d4726-3a2c-4e29-b6d8-1dc7139013a1-kube-api-access-mkbzb\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.613834 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-config-data\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.614386 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.616098 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176d4726-3a2c-4e29-b6d8-1dc7139013a1-logs\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.618296 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.620469 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-config-data\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.648613 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkbzb\" (UniqueName: \"kubernetes.io/projected/176d4726-3a2c-4e29-b6d8-1dc7139013a1-kube-api-access-mkbzb\") pod \"nova-metadata-0\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.672656 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.676758 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.681288 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.708062 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.715639 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.715687 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.715745 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.715793 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7swjd\" (UniqueName: \"kubernetes.io/projected/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-kube-api-access-7swjd\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.715823 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.715847 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-config\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.819676 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-config\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820127 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820187 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820215 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820241 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820303 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n28mm\" (UniqueName: \"kubernetes.io/projected/9a986167-8d84-4b52-91a1-5858e82b734f-kube-api-access-n28mm\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820329 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820373 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7swjd\" (UniqueName: \"kubernetes.io/projected/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-kube-api-access-7swjd\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.820399 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.821470 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.822108 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-config\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.822925 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.823521 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.825197 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.856412 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7swjd\" (UniqueName: \"kubernetes.io/projected/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-kube-api-access-7swjd\") pod \"dnsmasq-dns-845d6d6f59-759js\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.922078 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.922186 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.922254 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n28mm\" (UniqueName: \"kubernetes.io/projected/9a986167-8d84-4b52-91a1-5858e82b734f-kube-api-access-n28mm\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.937875 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.937942 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.943114 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n28mm\" (UniqueName: \"kubernetes.io/projected/9a986167-8d84-4b52-91a1-5858e82b734f-kube-api-access-n28mm\") pod \"nova-cell1-novncproxy-0\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.943620 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:44 crc kubenswrapper[5035]: I1002 09:47:44.968057 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.012520 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.234137 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.249386 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-7jwqw"] Oct 02 09:47:45 crc kubenswrapper[5035]: W1002 09:47:45.270770 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ac1551c_b8d0_4141_b334_2bcdc550cce8.slice/crio-f356943f551b79c7e56df1bd76eaa2a3fbcb2294df58b9bdd3548aa5d5fbff57 WatchSource:0}: Error finding container f356943f551b79c7e56df1bd76eaa2a3fbcb2294df58b9bdd3548aa5d5fbff57: Status 404 returned error can't find the container with id f356943f551b79c7e56df1bd76eaa2a3fbcb2294df58b9bdd3548aa5d5fbff57 Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.374851 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ac1551c-b8d0-4141-b334-2bcdc550cce8","Type":"ContainerStarted","Data":"f356943f551b79c7e56df1bd76eaa2a3fbcb2294df58b9bdd3548aa5d5fbff57"} Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.375777 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7jwqw" event={"ID":"104b87a6-20cd-459e-b095-eaa7de00ba31","Type":"ContainerStarted","Data":"cc4b3d82ca19acd7fe96ac1bba5d61c6ea06ac3e3cc28f22efebd2305c1eee3a"} Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.437077 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z7gpk"] Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.439002 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.441283 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.441843 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.450623 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z7gpk"] Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.482730 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:47:45 crc kubenswrapper[5035]: W1002 09:47:45.506223 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00e17dd1_f431_4b62_aaac_0f6d466870e4.slice/crio-f2c043054c253bc3ceca0475776a1e9b6f150d950cbbd05d577c835b50edd7fe WatchSource:0}: Error finding container f2c043054c253bc3ceca0475776a1e9b6f150d950cbbd05d577c835b50edd7fe: Status 404 returned error can't find the container with id f2c043054c253bc3ceca0475776a1e9b6f150d950cbbd05d577c835b50edd7fe Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.546168 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqz7n\" (UniqueName: \"kubernetes.io/projected/55b701fb-25b6-4e11-b46c-08a523c0d0ea-kube-api-access-qqz7n\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.546351 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-scripts\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.546388 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.546408 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-config-data\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.561560 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.629612 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-759js"] Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.638608 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.649142 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-scripts\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.649238 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.649268 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-config-data\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.649324 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqz7n\" (UniqueName: \"kubernetes.io/projected/55b701fb-25b6-4e11-b46c-08a523c0d0ea-kube-api-access-qqz7n\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.655057 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-config-data\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.656218 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.661446 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-scripts\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.675369 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqz7n\" (UniqueName: \"kubernetes.io/projected/55b701fb-25b6-4e11-b46c-08a523c0d0ea-kube-api-access-qqz7n\") pod \"nova-cell1-conductor-db-sync-z7gpk\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:45 crc kubenswrapper[5035]: I1002 09:47:45.763428 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:47:46 crc kubenswrapper[5035]: W1002 09:47:46.251057 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55b701fb_25b6_4e11_b46c_08a523c0d0ea.slice/crio-ff06e67a09ec5920509b604a378365981b07d2efc52c445e2dae3b0171088512 WatchSource:0}: Error finding container ff06e67a09ec5920509b604a378365981b07d2efc52c445e2dae3b0171088512: Status 404 returned error can't find the container with id ff06e67a09ec5920509b604a378365981b07d2efc52c445e2dae3b0171088512 Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.252902 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z7gpk"] Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.388386 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7jwqw" event={"ID":"104b87a6-20cd-459e-b095-eaa7de00ba31","Type":"ContainerStarted","Data":"083b2ea6a1503b549b21bb73e43ad1feecf7118e404dfcfbed8a661ed4df1bc4"} Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.389966 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"00e17dd1-f431-4b62-aaac-0f6d466870e4","Type":"ContainerStarted","Data":"f2c043054c253bc3ceca0475776a1e9b6f150d950cbbd05d577c835b50edd7fe"} Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.393316 5035 generic.go:334] "Generic (PLEG): container finished" podID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerID="49c6a0043f09fa0e68c4968b1ce55de721d9a1a073a4f8e976c48b283d70d338" exitCode=0 Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.393368 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-759js" event={"ID":"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c","Type":"ContainerDied","Data":"49c6a0043f09fa0e68c4968b1ce55de721d9a1a073a4f8e976c48b283d70d338"} Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.393416 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-759js" event={"ID":"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c","Type":"ContainerStarted","Data":"68ec136eee57ed7f57da5104f7c3cea8bd009915055a0d1a4de36f8a780e48d0"} Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.400954 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" event={"ID":"55b701fb-25b6-4e11-b46c-08a523c0d0ea","Type":"ContainerStarted","Data":"ff06e67a09ec5920509b604a378365981b07d2efc52c445e2dae3b0171088512"} Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.408747 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9a986167-8d84-4b52-91a1-5858e82b734f","Type":"ContainerStarted","Data":"c56c76315c0301a5b3f21073a4f357a887a5099092be29e53c8d68f212c7e77f"} Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.411094 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"176d4726-3a2c-4e29-b6d8-1dc7139013a1","Type":"ContainerStarted","Data":"3fb3fa535c5dd756a36070d74d76b894a6fdaa4e71af9f0c26d935a7fa708453"} Oct 02 09:47:46 crc kubenswrapper[5035]: I1002 09:47:46.413904 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-7jwqw" podStartSLOduration=2.413890061 podStartE2EDuration="2.413890061s" podCreationTimestamp="2025-10-02 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:47:46.404089566 +0000 UTC m=+1231.760433591" watchObservedRunningTime="2025-10-02 09:47:46.413890061 +0000 UTC m=+1231.770234086" Oct 02 09:47:47 crc kubenswrapper[5035]: I1002 09:47:47.423230 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" event={"ID":"55b701fb-25b6-4e11-b46c-08a523c0d0ea","Type":"ContainerStarted","Data":"e73d2ddf7b019bd706f80ab956f457a6e25470d37e54168a6c1f392c093fd74f"} Oct 02 09:47:47 crc kubenswrapper[5035]: I1002 09:47:47.452005 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" podStartSLOduration=2.451984509 podStartE2EDuration="2.451984509s" podCreationTimestamp="2025-10-02 09:47:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:47:47.439454494 +0000 UTC m=+1232.795798519" watchObservedRunningTime="2025-10-02 09:47:47.451984509 +0000 UTC m=+1232.808328534" Oct 02 09:47:47 crc kubenswrapper[5035]: I1002 09:47:47.452274 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-759js" event={"ID":"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c","Type":"ContainerStarted","Data":"2b3daca74aad2f3efdfff4772d258e1591628504c5da8ea5ab47a8bf55fa16aa"} Oct 02 09:47:47 crc kubenswrapper[5035]: I1002 09:47:47.452604 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:47 crc kubenswrapper[5035]: I1002 09:47:47.486124 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-759js" podStartSLOduration=3.486105113 podStartE2EDuration="3.486105113s" podCreationTimestamp="2025-10-02 09:47:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:47:47.48463471 +0000 UTC m=+1232.840978735" watchObservedRunningTime="2025-10-02 09:47:47.486105113 +0000 UTC m=+1232.842449138" Oct 02 09:47:47 crc kubenswrapper[5035]: I1002 09:47:47.968995 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:47 crc kubenswrapper[5035]: I1002 09:47:47.983570 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.511387 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9a986167-8d84-4b52-91a1-5858e82b734f","Type":"ContainerStarted","Data":"b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc"} Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.511630 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="9a986167-8d84-4b52-91a1-5858e82b734f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc" gracePeriod=30 Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.514701 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"176d4726-3a2c-4e29-b6d8-1dc7139013a1","Type":"ContainerStarted","Data":"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e"} Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.514768 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"176d4726-3a2c-4e29-b6d8-1dc7139013a1","Type":"ContainerStarted","Data":"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa"} Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.514794 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-log" containerID="cri-o://fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa" gracePeriod=30 Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.514867 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-metadata" containerID="cri-o://654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e" gracePeriod=30 Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.521112 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ac1551c-b8d0-4141-b334-2bcdc550cce8","Type":"ContainerStarted","Data":"7acba50cc2fab31193d50830e1e9a616634fd76cfda4cf0e71409016434234ab"} Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.521177 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ac1551c-b8d0-4141-b334-2bcdc550cce8","Type":"ContainerStarted","Data":"72d55129a09f7e84012d391268edcecfee46f4caca52baba9701adbf0836e6c2"} Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.526985 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"00e17dd1-f431-4b62-aaac-0f6d466870e4","Type":"ContainerStarted","Data":"756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93"} Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.552694 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.41701754 podStartE2EDuration="6.552672968s" podCreationTimestamp="2025-10-02 09:47:44 +0000 UTC" firstStartedPulling="2025-10-02 09:47:45.631009829 +0000 UTC m=+1230.987353854" lastFinishedPulling="2025-10-02 09:47:48.766665257 +0000 UTC m=+1234.123009282" observedRunningTime="2025-10-02 09:47:50.528855674 +0000 UTC m=+1235.885199699" watchObservedRunningTime="2025-10-02 09:47:50.552672968 +0000 UTC m=+1235.909016993" Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.563790 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.796090267 podStartE2EDuration="6.56375283s" podCreationTimestamp="2025-10-02 09:47:44 +0000 UTC" firstStartedPulling="2025-10-02 09:47:45.510339913 +0000 UTC m=+1230.866683938" lastFinishedPulling="2025-10-02 09:47:49.278002476 +0000 UTC m=+1234.634346501" observedRunningTime="2025-10-02 09:47:50.546198139 +0000 UTC m=+1235.902542204" watchObservedRunningTime="2025-10-02 09:47:50.56375283 +0000 UTC m=+1235.920096855" Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.579097 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.027683885 podStartE2EDuration="6.579073607s" podCreationTimestamp="2025-10-02 09:47:44 +0000 UTC" firstStartedPulling="2025-10-02 09:47:45.566982913 +0000 UTC m=+1230.923326938" lastFinishedPulling="2025-10-02 09:47:49.118372635 +0000 UTC m=+1234.474716660" observedRunningTime="2025-10-02 09:47:50.565833861 +0000 UTC m=+1235.922177896" watchObservedRunningTime="2025-10-02 09:47:50.579073607 +0000 UTC m=+1235.935417632" Oct 02 09:47:50 crc kubenswrapper[5035]: I1002 09:47:50.587475 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.742058203 podStartE2EDuration="6.58740291s" podCreationTimestamp="2025-10-02 09:47:44 +0000 UTC" firstStartedPulling="2025-10-02 09:47:45.272273486 +0000 UTC m=+1230.628617511" lastFinishedPulling="2025-10-02 09:47:49.117618193 +0000 UTC m=+1234.473962218" observedRunningTime="2025-10-02 09:47:50.582234919 +0000 UTC m=+1235.938578944" watchObservedRunningTime="2025-10-02 09:47:50.58740291 +0000 UTC m=+1235.943746935" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.114715 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.297720 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkbzb\" (UniqueName: \"kubernetes.io/projected/176d4726-3a2c-4e29-b6d8-1dc7139013a1-kube-api-access-mkbzb\") pod \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.298093 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-combined-ca-bundle\") pod \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.298148 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176d4726-3a2c-4e29-b6d8-1dc7139013a1-logs\") pod \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.298220 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-config-data\") pod \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\" (UID: \"176d4726-3a2c-4e29-b6d8-1dc7139013a1\") " Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.298592 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/176d4726-3a2c-4e29-b6d8-1dc7139013a1-logs" (OuterVolumeSpecName: "logs") pod "176d4726-3a2c-4e29-b6d8-1dc7139013a1" (UID: "176d4726-3a2c-4e29-b6d8-1dc7139013a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.299106 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/176d4726-3a2c-4e29-b6d8-1dc7139013a1-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.321879 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/176d4726-3a2c-4e29-b6d8-1dc7139013a1-kube-api-access-mkbzb" (OuterVolumeSpecName: "kube-api-access-mkbzb") pod "176d4726-3a2c-4e29-b6d8-1dc7139013a1" (UID: "176d4726-3a2c-4e29-b6d8-1dc7139013a1"). InnerVolumeSpecName "kube-api-access-mkbzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.348403 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-config-data" (OuterVolumeSpecName: "config-data") pod "176d4726-3a2c-4e29-b6d8-1dc7139013a1" (UID: "176d4726-3a2c-4e29-b6d8-1dc7139013a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.381799 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "176d4726-3a2c-4e29-b6d8-1dc7139013a1" (UID: "176d4726-3a2c-4e29-b6d8-1dc7139013a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.401106 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkbzb\" (UniqueName: \"kubernetes.io/projected/176d4726-3a2c-4e29-b6d8-1dc7139013a1-kube-api-access-mkbzb\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.401155 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.401169 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/176d4726-3a2c-4e29-b6d8-1dc7139013a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.539187 5035 generic.go:334] "Generic (PLEG): container finished" podID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerID="654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e" exitCode=0 Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.539227 5035 generic.go:334] "Generic (PLEG): container finished" podID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerID="fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa" exitCode=143 Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.540264 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"176d4726-3a2c-4e29-b6d8-1dc7139013a1","Type":"ContainerDied","Data":"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e"} Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.540298 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"176d4726-3a2c-4e29-b6d8-1dc7139013a1","Type":"ContainerDied","Data":"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa"} Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.540314 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"176d4726-3a2c-4e29-b6d8-1dc7139013a1","Type":"ContainerDied","Data":"3fb3fa535c5dd756a36070d74d76b894a6fdaa4e71af9f0c26d935a7fa708453"} Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.540336 5035 scope.go:117] "RemoveContainer" containerID="654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.540691 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.569354 5035 scope.go:117] "RemoveContainer" containerID="fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.583367 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.596180 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.604467 5035 scope.go:117] "RemoveContainer" containerID="654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e" Oct 02 09:47:51 crc kubenswrapper[5035]: E1002 09:47:51.604995 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e\": container with ID starting with 654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e not found: ID does not exist" containerID="654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.605040 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e"} err="failed to get container status \"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e\": rpc error: code = NotFound desc = could not find container \"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e\": container with ID starting with 654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e not found: ID does not exist" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.605070 5035 scope.go:117] "RemoveContainer" containerID="fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa" Oct 02 09:47:51 crc kubenswrapper[5035]: E1002 09:47:51.606312 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa\": container with ID starting with fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa not found: ID does not exist" containerID="fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.606374 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa"} err="failed to get container status \"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa\": rpc error: code = NotFound desc = could not find container \"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa\": container with ID starting with fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa not found: ID does not exist" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.606394 5035 scope.go:117] "RemoveContainer" containerID="654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.607027 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e"} err="failed to get container status \"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e\": rpc error: code = NotFound desc = could not find container \"654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e\": container with ID starting with 654fe406baab1ae6d9791306a9a09086363ce7ab2313781074dc7b52eca9ec4e not found: ID does not exist" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.607068 5035 scope.go:117] "RemoveContainer" containerID="fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.607327 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa"} err="failed to get container status \"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa\": rpc error: code = NotFound desc = could not find container \"fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa\": container with ID starting with fbcf5a329ba42f9100aaa74a51a1235cbc7e9125b89d5942b1c84168d3939aaa not found: ID does not exist" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.608614 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:51 crc kubenswrapper[5035]: E1002 09:47:51.609121 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-metadata" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.609143 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-metadata" Oct 02 09:47:51 crc kubenswrapper[5035]: E1002 09:47:51.609176 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-log" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.609186 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-log" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.609374 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-log" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.609397 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" containerName="nova-metadata-metadata" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.610752 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.613451 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.613896 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.616971 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.708781 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-logs\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.708836 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-config-data\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.709005 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.709169 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpndz\" (UniqueName: \"kubernetes.io/projected/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-kube-api-access-hpndz\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.709224 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.811256 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.811345 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpndz\" (UniqueName: \"kubernetes.io/projected/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-kube-api-access-hpndz\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.811378 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.811479 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-logs\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.811512 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-config-data\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.812738 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-logs\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.815642 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-config-data\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.815648 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.817982 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.835263 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpndz\" (UniqueName: \"kubernetes.io/projected/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-kube-api-access-hpndz\") pod \"nova-metadata-0\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " pod="openstack/nova-metadata-0" Oct 02 09:47:51 crc kubenswrapper[5035]: I1002 09:47:51.939782 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:52 crc kubenswrapper[5035]: I1002 09:47:52.177678 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="176d4726-3a2c-4e29-b6d8-1dc7139013a1" path="/var/lib/kubelet/pods/176d4726-3a2c-4e29-b6d8-1dc7139013a1/volumes" Oct 02 09:47:52 crc kubenswrapper[5035]: I1002 09:47:52.386918 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:52 crc kubenswrapper[5035]: W1002 09:47:52.424517 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfaa548b2_2ce1_4af6_bb43_6c5a4ec0df23.slice/crio-d15c8ca73883eef455d5fffaa7b1c91a49e41ad34636f1d60a18f85b45c9abee WatchSource:0}: Error finding container d15c8ca73883eef455d5fffaa7b1c91a49e41ad34636f1d60a18f85b45c9abee: Status 404 returned error can't find the container with id d15c8ca73883eef455d5fffaa7b1c91a49e41ad34636f1d60a18f85b45c9abee Oct 02 09:47:52 crc kubenswrapper[5035]: I1002 09:47:52.567344 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23","Type":"ContainerStarted","Data":"d15c8ca73883eef455d5fffaa7b1c91a49e41ad34636f1d60a18f85b45c9abee"} Oct 02 09:47:53 crc kubenswrapper[5035]: I1002 09:47:53.579371 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23","Type":"ContainerStarted","Data":"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154"} Oct 02 09:47:53 crc kubenswrapper[5035]: I1002 09:47:53.579849 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23","Type":"ContainerStarted","Data":"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b"} Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.531181 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.531237 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.593114 5035 generic.go:334] "Generic (PLEG): container finished" podID="104b87a6-20cd-459e-b095-eaa7de00ba31" containerID="083b2ea6a1503b549b21bb73e43ad1feecf7118e404dfcfbed8a661ed4df1bc4" exitCode=0 Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.593365 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7jwqw" event={"ID":"104b87a6-20cd-459e-b095-eaa7de00ba31","Type":"ContainerDied","Data":"083b2ea6a1503b549b21bb73e43ad1feecf7118e404dfcfbed8a661ed4df1bc4"} Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.616475 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.616750 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.625282 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.625260296 podStartE2EDuration="3.625260296s" podCreationTimestamp="2025-10-02 09:47:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:47:53.605300346 +0000 UTC m=+1238.961644371" watchObservedRunningTime="2025-10-02 09:47:54.625260296 +0000 UTC m=+1239.981604331" Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.732971 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 09:47:54 crc kubenswrapper[5035]: I1002 09:47:54.970803 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.013841 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.038375 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-4f72t"] Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.038660 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" podUID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerName="dnsmasq-dns" containerID="cri-o://a10d5e846c8ea3f7112b3a156244cc8d64d57eeb3b7d582c28802531b8587bbe" gracePeriod=10 Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.607817 5035 generic.go:334] "Generic (PLEG): container finished" podID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerID="a10d5e846c8ea3f7112b3a156244cc8d64d57eeb3b7d582c28802531b8587bbe" exitCode=0 Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.608064 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" event={"ID":"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd","Type":"ContainerDied","Data":"a10d5e846c8ea3f7112b3a156244cc8d64d57eeb3b7d582c28802531b8587bbe"} Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.608176 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" event={"ID":"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd","Type":"ContainerDied","Data":"a95b8608cdd74b27410b68eefe1a5f18163559361d0cbd547718084a33011303"} Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.608197 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a95b8608cdd74b27410b68eefe1a5f18163559361d0cbd547718084a33011303" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.619446 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.623234 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.197:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.643965 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.649765 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.810682 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-config\") pod \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.810805 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-sb\") pod \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.810844 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-svc\") pod \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.810881 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgrvz\" (UniqueName: \"kubernetes.io/projected/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-kube-api-access-zgrvz\") pod \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.810945 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-nb\") pod \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.810996 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-swift-storage-0\") pod \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\" (UID: \"f1fa0550-e42e-4eaf-8437-7cd24d12e5cd\") " Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.843423 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-kube-api-access-zgrvz" (OuterVolumeSpecName: "kube-api-access-zgrvz") pod "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" (UID: "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd"). InnerVolumeSpecName "kube-api-access-zgrvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.914169 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgrvz\" (UniqueName: \"kubernetes.io/projected/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-kube-api-access-zgrvz\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.952079 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-config" (OuterVolumeSpecName: "config") pod "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" (UID: "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.963322 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" (UID: "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.965986 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" (UID: "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:47:55 crc kubenswrapper[5035]: I1002 09:47:55.979968 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" (UID: "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.001785 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" (UID: "f1fa0550-e42e-4eaf-8437-7cd24d12e5cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.017945 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.018085 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.018163 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.018228 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.018284 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.030369 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.119500 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-scripts\") pod \"104b87a6-20cd-459e-b095-eaa7de00ba31\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.119751 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9wps\" (UniqueName: \"kubernetes.io/projected/104b87a6-20cd-459e-b095-eaa7de00ba31-kube-api-access-h9wps\") pod \"104b87a6-20cd-459e-b095-eaa7de00ba31\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.119840 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-config-data\") pod \"104b87a6-20cd-459e-b095-eaa7de00ba31\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.119932 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-combined-ca-bundle\") pod \"104b87a6-20cd-459e-b095-eaa7de00ba31\" (UID: \"104b87a6-20cd-459e-b095-eaa7de00ba31\") " Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.126474 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-scripts" (OuterVolumeSpecName: "scripts") pod "104b87a6-20cd-459e-b095-eaa7de00ba31" (UID: "104b87a6-20cd-459e-b095-eaa7de00ba31"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.126482 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/104b87a6-20cd-459e-b095-eaa7de00ba31-kube-api-access-h9wps" (OuterVolumeSpecName: "kube-api-access-h9wps") pod "104b87a6-20cd-459e-b095-eaa7de00ba31" (UID: "104b87a6-20cd-459e-b095-eaa7de00ba31"). InnerVolumeSpecName "kube-api-access-h9wps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.154984 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "104b87a6-20cd-459e-b095-eaa7de00ba31" (UID: "104b87a6-20cd-459e-b095-eaa7de00ba31"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.162563 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-config-data" (OuterVolumeSpecName: "config-data") pod "104b87a6-20cd-459e-b095-eaa7de00ba31" (UID: "104b87a6-20cd-459e-b095-eaa7de00ba31"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.222161 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.222215 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.222225 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9wps\" (UniqueName: \"kubernetes.io/projected/104b87a6-20cd-459e-b095-eaa7de00ba31-kube-api-access-h9wps\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.222236 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/104b87a6-20cd-459e-b095-eaa7de00ba31-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.618876 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-4f72t" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.619128 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-7jwqw" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.619153 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-7jwqw" event={"ID":"104b87a6-20cd-459e-b095-eaa7de00ba31","Type":"ContainerDied","Data":"cc4b3d82ca19acd7fe96ac1bba5d61c6ea06ac3e3cc28f22efebd2305c1eee3a"} Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.619359 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc4b3d82ca19acd7fe96ac1bba5d61c6ea06ac3e3cc28f22efebd2305c1eee3a" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.643727 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-4f72t"] Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.658980 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-4f72t"] Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.793056 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.793369 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-log" containerID="cri-o://72d55129a09f7e84012d391268edcecfee46f4caca52baba9701adbf0836e6c2" gracePeriod=30 Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.793439 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-api" containerID="cri-o://7acba50cc2fab31193d50830e1e9a616634fd76cfda4cf0e71409016434234ab" gracePeriod=30 Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.805815 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.821555 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.821803 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-log" containerID="cri-o://483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b" gracePeriod=30 Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.821856 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-metadata" containerID="cri-o://f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154" gracePeriod=30 Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.940146 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 09:47:56 crc kubenswrapper[5035]: I1002 09:47:56.940196 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.499505 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.649810 5035 generic.go:334] "Generic (PLEG): container finished" podID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerID="72d55129a09f7e84012d391268edcecfee46f4caca52baba9701adbf0836e6c2" exitCode=143 Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.649906 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ac1551c-b8d0-4141-b334-2bcdc550cce8","Type":"ContainerDied","Data":"72d55129a09f7e84012d391268edcecfee46f4caca52baba9701adbf0836e6c2"} Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.654020 5035 generic.go:334] "Generic (PLEG): container finished" podID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerID="f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154" exitCode=0 Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.654049 5035 generic.go:334] "Generic (PLEG): container finished" podID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerID="483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b" exitCode=143 Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.654677 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.654771 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23","Type":"ContainerDied","Data":"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154"} Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.654796 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23","Type":"ContainerDied","Data":"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b"} Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.654807 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23","Type":"ContainerDied","Data":"d15c8ca73883eef455d5fffaa7b1c91a49e41ad34636f1d60a18f85b45c9abee"} Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.654822 5035 scope.go:117] "RemoveContainer" containerID="f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.655344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-combined-ca-bundle\") pod \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.655430 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-nova-metadata-tls-certs\") pod \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.655495 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-config-data\") pod \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.655589 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-logs\") pod \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.655687 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpndz\" (UniqueName: \"kubernetes.io/projected/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-kube-api-access-hpndz\") pod \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\" (UID: \"faa548b2-2ce1-4af6-bb43-6c5a4ec0df23\") " Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.658023 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-logs" (OuterVolumeSpecName: "logs") pod "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" (UID: "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.662025 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-kube-api-access-hpndz" (OuterVolumeSpecName: "kube-api-access-hpndz") pod "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" (UID: "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23"). InnerVolumeSpecName "kube-api-access-hpndz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.708741 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-config-data" (OuterVolumeSpecName: "config-data") pod "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" (UID: "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.751660 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" (UID: "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.757658 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.757687 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpndz\" (UniqueName: \"kubernetes.io/projected/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-kube-api-access-hpndz\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.757697 5035 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.757705 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.769959 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" (UID: "faa548b2-2ce1-4af6-bb43-6c5a4ec0df23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.858857 5035 scope.go:117] "RemoveContainer" containerID="483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.860128 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.876645 5035 scope.go:117] "RemoveContainer" containerID="f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154" Oct 02 09:47:57 crc kubenswrapper[5035]: E1002 09:47:57.877166 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154\": container with ID starting with f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154 not found: ID does not exist" containerID="f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.877202 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154"} err="failed to get container status \"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154\": rpc error: code = NotFound desc = could not find container \"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154\": container with ID starting with f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154 not found: ID does not exist" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.877229 5035 scope.go:117] "RemoveContainer" containerID="483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b" Oct 02 09:47:57 crc kubenswrapper[5035]: E1002 09:47:57.877458 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b\": container with ID starting with 483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b not found: ID does not exist" containerID="483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.877493 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b"} err="failed to get container status \"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b\": rpc error: code = NotFound desc = could not find container \"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b\": container with ID starting with 483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b not found: ID does not exist" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.877511 5035 scope.go:117] "RemoveContainer" containerID="f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.877898 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154"} err="failed to get container status \"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154\": rpc error: code = NotFound desc = could not find container \"f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154\": container with ID starting with f06f688f66007cb3f2d7cc214c5bca2985c3a26080de433e92b73be801539154 not found: ID does not exist" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.877921 5035 scope.go:117] "RemoveContainer" containerID="483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b" Oct 02 09:47:57 crc kubenswrapper[5035]: I1002 09:47:57.878110 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b"} err="failed to get container status \"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b\": rpc error: code = NotFound desc = could not find container \"483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b\": container with ID starting with 483f3c8cf26914f9cc9aa8b4744b467387921a412264acbe91eb1e3c66d7b97b not found: ID does not exist" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.027040 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.035283 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057118 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:58 crc kubenswrapper[5035]: E1002 09:47:58.057581 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerName="init" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057600 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerName="init" Oct 02 09:47:58 crc kubenswrapper[5035]: E1002 09:47:58.057620 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-metadata" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057628 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-metadata" Oct 02 09:47:58 crc kubenswrapper[5035]: E1002 09:47:58.057640 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-log" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057647 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-log" Oct 02 09:47:58 crc kubenswrapper[5035]: E1002 09:47:58.057670 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="104b87a6-20cd-459e-b095-eaa7de00ba31" containerName="nova-manage" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057678 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="104b87a6-20cd-459e-b095-eaa7de00ba31" containerName="nova-manage" Oct 02 09:47:58 crc kubenswrapper[5035]: E1002 09:47:58.057689 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerName="dnsmasq-dns" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057708 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerName="dnsmasq-dns" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057884 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-log" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057902 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" containerName="dnsmasq-dns" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057914 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="104b87a6-20cd-459e-b095-eaa7de00ba31" containerName="nova-manage" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.057926 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" containerName="nova-metadata-metadata" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.058995 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.065208 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.065641 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.074955 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.167936 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.168311 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-config-data\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.168335 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57a6e9b-12ef-4611-861f-f57215d0edc6-logs\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.168359 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.168422 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7rh7\" (UniqueName: \"kubernetes.io/projected/e57a6e9b-12ef-4611-861f-f57215d0edc6-kube-api-access-g7rh7\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.176421 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1fa0550-e42e-4eaf-8437-7cd24d12e5cd" path="/var/lib/kubelet/pods/f1fa0550-e42e-4eaf-8437-7cd24d12e5cd/volumes" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.177052 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faa548b2-2ce1-4af6-bb43-6c5a4ec0df23" path="/var/lib/kubelet/pods/faa548b2-2ce1-4af6-bb43-6c5a4ec0df23/volumes" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.269808 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.269921 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-config-data\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.269955 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57a6e9b-12ef-4611-861f-f57215d0edc6-logs\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.269981 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.270053 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7rh7\" (UniqueName: \"kubernetes.io/projected/e57a6e9b-12ef-4611-861f-f57215d0edc6-kube-api-access-g7rh7\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.270704 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57a6e9b-12ef-4611-861f-f57215d0edc6-logs\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.273663 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.274127 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-config-data\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.275164 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.318377 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7rh7\" (UniqueName: \"kubernetes.io/projected/e57a6e9b-12ef-4611-861f-f57215d0edc6-kube-api-access-g7rh7\") pod \"nova-metadata-0\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.376676 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.665602 5035 generic.go:334] "Generic (PLEG): container finished" podID="55b701fb-25b6-4e11-b46c-08a523c0d0ea" containerID="e73d2ddf7b019bd706f80ab956f457a6e25470d37e54168a6c1f392c093fd74f" exitCode=0 Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.665688 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" event={"ID":"55b701fb-25b6-4e11-b46c-08a523c0d0ea","Type":"ContainerDied","Data":"e73d2ddf7b019bd706f80ab956f457a6e25470d37e54168a6c1f392c093fd74f"} Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.668618 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="00e17dd1-f431-4b62-aaac-0f6d466870e4" containerName="nova-scheduler-scheduler" containerID="cri-o://756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" gracePeriod=30 Oct 02 09:47:58 crc kubenswrapper[5035]: I1002 09:47:58.844424 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:47:59 crc kubenswrapper[5035]: E1002 09:47:59.617916 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 09:47:59 crc kubenswrapper[5035]: E1002 09:47:59.619578 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 09:47:59 crc kubenswrapper[5035]: E1002 09:47:59.621135 5035 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 09:47:59 crc kubenswrapper[5035]: E1002 09:47:59.621181 5035 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="00e17dd1-f431-4b62-aaac-0f6d466870e4" containerName="nova-scheduler-scheduler" Oct 02 09:47:59 crc kubenswrapper[5035]: I1002 09:47:59.680144 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e57a6e9b-12ef-4611-861f-f57215d0edc6","Type":"ContainerStarted","Data":"0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659"} Oct 02 09:47:59 crc kubenswrapper[5035]: I1002 09:47:59.680200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e57a6e9b-12ef-4611-861f-f57215d0edc6","Type":"ContainerStarted","Data":"3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d"} Oct 02 09:47:59 crc kubenswrapper[5035]: I1002 09:47:59.680214 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e57a6e9b-12ef-4611-861f-f57215d0edc6","Type":"ContainerStarted","Data":"dc44c589879a4b332744650db2625730f5912abc0444ab66d1453ab531615eca"} Oct 02 09:47:59 crc kubenswrapper[5035]: I1002 09:47:59.705511 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.705490936 podStartE2EDuration="1.705490936s" podCreationTimestamp="2025-10-02 09:47:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:47:59.6984272 +0000 UTC m=+1245.054771225" watchObservedRunningTime="2025-10-02 09:47:59.705490936 +0000 UTC m=+1245.061834961" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.060820 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.207690 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-scripts\") pod \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.207811 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-combined-ca-bundle\") pod \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.207913 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-config-data\") pod \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.208114 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqz7n\" (UniqueName: \"kubernetes.io/projected/55b701fb-25b6-4e11-b46c-08a523c0d0ea-kube-api-access-qqz7n\") pod \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\" (UID: \"55b701fb-25b6-4e11-b46c-08a523c0d0ea\") " Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.224557 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-scripts" (OuterVolumeSpecName: "scripts") pod "55b701fb-25b6-4e11-b46c-08a523c0d0ea" (UID: "55b701fb-25b6-4e11-b46c-08a523c0d0ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.224591 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b701fb-25b6-4e11-b46c-08a523c0d0ea-kube-api-access-qqz7n" (OuterVolumeSpecName: "kube-api-access-qqz7n") pod "55b701fb-25b6-4e11-b46c-08a523c0d0ea" (UID: "55b701fb-25b6-4e11-b46c-08a523c0d0ea"). InnerVolumeSpecName "kube-api-access-qqz7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.239348 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55b701fb-25b6-4e11-b46c-08a523c0d0ea" (UID: "55b701fb-25b6-4e11-b46c-08a523c0d0ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.251119 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-config-data" (OuterVolumeSpecName: "config-data") pod "55b701fb-25b6-4e11-b46c-08a523c0d0ea" (UID: "55b701fb-25b6-4e11-b46c-08a523c0d0ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.311310 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.311333 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.311346 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b701fb-25b6-4e11-b46c-08a523c0d0ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.311355 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqz7n\" (UniqueName: \"kubernetes.io/projected/55b701fb-25b6-4e11-b46c-08a523c0d0ea-kube-api-access-qqz7n\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.690494 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" event={"ID":"55b701fb-25b6-4e11-b46c-08a523c0d0ea","Type":"ContainerDied","Data":"ff06e67a09ec5920509b604a378365981b07d2efc52c445e2dae3b0171088512"} Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.690870 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff06e67a09ec5920509b604a378365981b07d2efc52c445e2dae3b0171088512" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.690518 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-z7gpk" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.754518 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 09:48:00 crc kubenswrapper[5035]: E1002 09:48:00.754975 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b701fb-25b6-4e11-b46c-08a523c0d0ea" containerName="nova-cell1-conductor-db-sync" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.754992 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b701fb-25b6-4e11-b46c-08a523c0d0ea" containerName="nova-cell1-conductor-db-sync" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.755183 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b701fb-25b6-4e11-b46c-08a523c0d0ea" containerName="nova-cell1-conductor-db-sync" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.756548 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.758441 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.777592 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.821680 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swl5r\" (UniqueName: \"kubernetes.io/projected/ad2f225f-00ca-455a-8061-e45abdb56864-kube-api-access-swl5r\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.821758 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2f225f-00ca-455a-8061-e45abdb56864-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.821814 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad2f225f-00ca-455a-8061-e45abdb56864-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.923596 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swl5r\" (UniqueName: \"kubernetes.io/projected/ad2f225f-00ca-455a-8061-e45abdb56864-kube-api-access-swl5r\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.923958 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2f225f-00ca-455a-8061-e45abdb56864-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.924095 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad2f225f-00ca-455a-8061-e45abdb56864-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.927907 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ad2f225f-00ca-455a-8061-e45abdb56864-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.929969 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ad2f225f-00ca-455a-8061-e45abdb56864-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:00 crc kubenswrapper[5035]: I1002 09:48:00.939401 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swl5r\" (UniqueName: \"kubernetes.io/projected/ad2f225f-00ca-455a-8061-e45abdb56864-kube-api-access-swl5r\") pod \"nova-cell1-conductor-0\" (UID: \"ad2f225f-00ca-455a-8061-e45abdb56864\") " pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:01 crc kubenswrapper[5035]: I1002 09:48:01.086215 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:01 crc kubenswrapper[5035]: I1002 09:48:01.589053 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 09:48:01 crc kubenswrapper[5035]: I1002 09:48:01.702850 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ad2f225f-00ca-455a-8061-e45abdb56864","Type":"ContainerStarted","Data":"3cf966552a89646da2ee7695f8eb8f67a2b3c33774f3fe3b07dec18d35b2bc82"} Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.374108 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.472646 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnntj\" (UniqueName: \"kubernetes.io/projected/00e17dd1-f431-4b62-aaac-0f6d466870e4-kube-api-access-cnntj\") pod \"00e17dd1-f431-4b62-aaac-0f6d466870e4\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.473228 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-combined-ca-bundle\") pod \"00e17dd1-f431-4b62-aaac-0f6d466870e4\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.473282 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-config-data\") pod \"00e17dd1-f431-4b62-aaac-0f6d466870e4\" (UID: \"00e17dd1-f431-4b62-aaac-0f6d466870e4\") " Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.513955 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00e17dd1-f431-4b62-aaac-0f6d466870e4-kube-api-access-cnntj" (OuterVolumeSpecName: "kube-api-access-cnntj") pod "00e17dd1-f431-4b62-aaac-0f6d466870e4" (UID: "00e17dd1-f431-4b62-aaac-0f6d466870e4"). InnerVolumeSpecName "kube-api-access-cnntj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.546006 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-config-data" (OuterVolumeSpecName: "config-data") pod "00e17dd1-f431-4b62-aaac-0f6d466870e4" (UID: "00e17dd1-f431-4b62-aaac-0f6d466870e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.546035 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00e17dd1-f431-4b62-aaac-0f6d466870e4" (UID: "00e17dd1-f431-4b62-aaac-0f6d466870e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.576097 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.576132 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnntj\" (UniqueName: \"kubernetes.io/projected/00e17dd1-f431-4b62-aaac-0f6d466870e4-kube-api-access-cnntj\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.576148 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00e17dd1-f431-4b62-aaac-0f6d466870e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.719644 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ad2f225f-00ca-455a-8061-e45abdb56864","Type":"ContainerStarted","Data":"5ff58215de875e8f0788a8ba6d8ceee0ddee5cac4219048b9a27c5121196cda9"} Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.719809 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.722192 5035 generic.go:334] "Generic (PLEG): container finished" podID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerID="7acba50cc2fab31193d50830e1e9a616634fd76cfda4cf0e71409016434234ab" exitCode=0 Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.722286 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ac1551c-b8d0-4141-b334-2bcdc550cce8","Type":"ContainerDied","Data":"7acba50cc2fab31193d50830e1e9a616634fd76cfda4cf0e71409016434234ab"} Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.724317 5035 generic.go:334] "Generic (PLEG): container finished" podID="00e17dd1-f431-4b62-aaac-0f6d466870e4" containerID="756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" exitCode=0 Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.724353 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"00e17dd1-f431-4b62-aaac-0f6d466870e4","Type":"ContainerDied","Data":"756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93"} Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.724374 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"00e17dd1-f431-4b62-aaac-0f6d466870e4","Type":"ContainerDied","Data":"f2c043054c253bc3ceca0475776a1e9b6f150d950cbbd05d577c835b50edd7fe"} Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.724394 5035 scope.go:117] "RemoveContainer" containerID="756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.724519 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.744326 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.744306192 podStartE2EDuration="2.744306192s" podCreationTimestamp="2025-10-02 09:48:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:02.736474714 +0000 UTC m=+1248.092818739" watchObservedRunningTime="2025-10-02 09:48:02.744306192 +0000 UTC m=+1248.100650217" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.745621 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.765730 5035 scope.go:117] "RemoveContainer" containerID="756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" Oct 02 09:48:02 crc kubenswrapper[5035]: E1002 09:48:02.766159 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93\": container with ID starting with 756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93 not found: ID does not exist" containerID="756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.766188 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93"} err="failed to get container status \"756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93\": rpc error: code = NotFound desc = could not find container \"756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93\": container with ID starting with 756cbabf09af28a7643029805e4447b735f73fb4143caa97ac51a4fdbcda3c93 not found: ID does not exist" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.776608 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.778480 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-combined-ca-bundle\") pod \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.778584 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac1551c-b8d0-4141-b334-2bcdc550cce8-logs\") pod \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.778606 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-config-data\") pod \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.778706 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2d2t\" (UniqueName: \"kubernetes.io/projected/6ac1551c-b8d0-4141-b334-2bcdc550cce8-kube-api-access-c2d2t\") pod \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\" (UID: \"6ac1551c-b8d0-4141-b334-2bcdc550cce8\") " Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.779025 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ac1551c-b8d0-4141-b334-2bcdc550cce8-logs" (OuterVolumeSpecName: "logs") pod "6ac1551c-b8d0-4141-b334-2bcdc550cce8" (UID: "6ac1551c-b8d0-4141-b334-2bcdc550cce8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.779433 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ac1551c-b8d0-4141-b334-2bcdc550cce8-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.791410 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.796911 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac1551c-b8d0-4141-b334-2bcdc550cce8-kube-api-access-c2d2t" (OuterVolumeSpecName: "kube-api-access-c2d2t") pod "6ac1551c-b8d0-4141-b334-2bcdc550cce8" (UID: "6ac1551c-b8d0-4141-b334-2bcdc550cce8"). InnerVolumeSpecName "kube-api-access-c2d2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.805104 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:02 crc kubenswrapper[5035]: E1002 09:48:02.805547 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-log" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.805561 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-log" Oct 02 09:48:02 crc kubenswrapper[5035]: E1002 09:48:02.805585 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00e17dd1-f431-4b62-aaac-0f6d466870e4" containerName="nova-scheduler-scheduler" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.805591 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="00e17dd1-f431-4b62-aaac-0f6d466870e4" containerName="nova-scheduler-scheduler" Oct 02 09:48:02 crc kubenswrapper[5035]: E1002 09:48:02.805604 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-api" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.805611 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-api" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.805798 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-log" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.805826 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="00e17dd1-f431-4b62-aaac-0f6d466870e4" containerName="nova-scheduler-scheduler" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.805841 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" containerName="nova-api-api" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.806449 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.808587 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.810819 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-config-data" (OuterVolumeSpecName: "config-data") pod "6ac1551c-b8d0-4141-b334-2bcdc550cce8" (UID: "6ac1551c-b8d0-4141-b334-2bcdc550cce8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.817086 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.824930 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ac1551c-b8d0-4141-b334-2bcdc550cce8" (UID: "6ac1551c-b8d0-4141-b334-2bcdc550cce8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.886074 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mnnr\" (UniqueName: \"kubernetes.io/projected/4d1e4714-45d9-4930-a186-43210468b573-kube-api-access-7mnnr\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.886136 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.886273 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-config-data\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.886348 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2d2t\" (UniqueName: \"kubernetes.io/projected/6ac1551c-b8d0-4141-b334-2bcdc550cce8-kube-api-access-c2d2t\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.886360 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.886369 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ac1551c-b8d0-4141-b334-2bcdc550cce8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.988139 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-config-data\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.988360 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mnnr\" (UniqueName: \"kubernetes.io/projected/4d1e4714-45d9-4930-a186-43210468b573-kube-api-access-7mnnr\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.991696 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.994371 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:02 crc kubenswrapper[5035]: I1002 09:48:02.994775 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-config-data\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.014699 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mnnr\" (UniqueName: \"kubernetes.io/projected/4d1e4714-45d9-4930-a186-43210468b573-kube-api-access-7mnnr\") pod \"nova-scheduler-0\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.132723 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.377653 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.377708 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.584753 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:03 crc kubenswrapper[5035]: W1002 09:48:03.608604 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d1e4714_45d9_4930_a186_43210468b573.slice/crio-22485c98714e4081d54ed8c66ec8a44be9958a7ed39b3f8cc2085f7242ba62b9 WatchSource:0}: Error finding container 22485c98714e4081d54ed8c66ec8a44be9958a7ed39b3f8cc2085f7242ba62b9: Status 404 returned error can't find the container with id 22485c98714e4081d54ed8c66ec8a44be9958a7ed39b3f8cc2085f7242ba62b9 Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.737487 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6ac1551c-b8d0-4141-b334-2bcdc550cce8","Type":"ContainerDied","Data":"f356943f551b79c7e56df1bd76eaa2a3fbcb2294df58b9bdd3548aa5d5fbff57"} Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.737608 5035 scope.go:117] "RemoveContainer" containerID="7acba50cc2fab31193d50830e1e9a616634fd76cfda4cf0e71409016434234ab" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.737732 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.749983 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d1e4714-45d9-4930-a186-43210468b573","Type":"ContainerStarted","Data":"22485c98714e4081d54ed8c66ec8a44be9958a7ed39b3f8cc2085f7242ba62b9"} Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.787036 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.790180 5035 scope.go:117] "RemoveContainer" containerID="72d55129a09f7e84012d391268edcecfee46f4caca52baba9701adbf0836e6c2" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.810282 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.843297 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.845257 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.848206 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.853980 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.909097 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdd8m\" (UniqueName: \"kubernetes.io/projected/27578bd8-1ea0-4dd8-bef6-c1cd69815419-kube-api-access-mdd8m\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.909400 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.909615 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-config-data\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:03 crc kubenswrapper[5035]: I1002 09:48:03.909802 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27578bd8-1ea0-4dd8-bef6-c1cd69815419-logs\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.012576 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.012932 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdd8m\" (UniqueName: \"kubernetes.io/projected/27578bd8-1ea0-4dd8-bef6-c1cd69815419-kube-api-access-mdd8m\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.013241 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-config-data\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.013498 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27578bd8-1ea0-4dd8-bef6-c1cd69815419-logs\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.013854 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27578bd8-1ea0-4dd8-bef6-c1cd69815419-logs\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.017964 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-config-data\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.019029 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.034009 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdd8m\" (UniqueName: \"kubernetes.io/projected/27578bd8-1ea0-4dd8-bef6-c1cd69815419-kube-api-access-mdd8m\") pod \"nova-api-0\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.166067 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.176202 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00e17dd1-f431-4b62-aaac-0f6d466870e4" path="/var/lib/kubelet/pods/00e17dd1-f431-4b62-aaac-0f6d466870e4/volumes" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.177054 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac1551c-b8d0-4141-b334-2bcdc550cce8" path="/var/lib/kubelet/pods/6ac1551c-b8d0-4141-b334-2bcdc550cce8/volumes" Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.684146 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.765617 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d1e4714-45d9-4930-a186-43210468b573","Type":"ContainerStarted","Data":"ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5"} Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.771723 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27578bd8-1ea0-4dd8-bef6-c1cd69815419","Type":"ContainerStarted","Data":"aa9f98e434c3f7df8f9e65b9f2e620007aa00da43e5135d8698073bbbd8d156d"} Oct 02 09:48:04 crc kubenswrapper[5035]: I1002 09:48:04.798568 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.798547909 podStartE2EDuration="2.798547909s" podCreationTimestamp="2025-10-02 09:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:04.785243801 +0000 UTC m=+1250.141587836" watchObservedRunningTime="2025-10-02 09:48:04.798547909 +0000 UTC m=+1250.154891934" Oct 02 09:48:05 crc kubenswrapper[5035]: I1002 09:48:05.791437 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27578bd8-1ea0-4dd8-bef6-c1cd69815419","Type":"ContainerStarted","Data":"6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76"} Oct 02 09:48:05 crc kubenswrapper[5035]: I1002 09:48:05.791757 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27578bd8-1ea0-4dd8-bef6-c1cd69815419","Type":"ContainerStarted","Data":"0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc"} Oct 02 09:48:05 crc kubenswrapper[5035]: I1002 09:48:05.815629 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.815606205 podStartE2EDuration="2.815606205s" podCreationTimestamp="2025-10-02 09:48:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:05.809292061 +0000 UTC m=+1251.165636106" watchObservedRunningTime="2025-10-02 09:48:05.815606205 +0000 UTC m=+1251.171950240" Oct 02 09:48:06 crc kubenswrapper[5035]: I1002 09:48:06.114986 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 09:48:06 crc kubenswrapper[5035]: I1002 09:48:06.984585 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 09:48:08 crc kubenswrapper[5035]: I1002 09:48:08.134664 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 09:48:08 crc kubenswrapper[5035]: I1002 09:48:08.377298 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 09:48:08 crc kubenswrapper[5035]: I1002 09:48:08.377350 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 09:48:09 crc kubenswrapper[5035]: I1002 09:48:09.389739 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 09:48:09 crc kubenswrapper[5035]: I1002 09:48:09.389776 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 09:48:13 crc kubenswrapper[5035]: I1002 09:48:13.134424 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 09:48:13 crc kubenswrapper[5035]: I1002 09:48:13.178577 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 09:48:13 crc kubenswrapper[5035]: I1002 09:48:13.896313 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 09:48:14 crc kubenswrapper[5035]: I1002 09:48:14.175280 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 09:48:14 crc kubenswrapper[5035]: I1002 09:48:14.175331 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 09:48:15 crc kubenswrapper[5035]: I1002 09:48:15.249785 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:48:15 crc kubenswrapper[5035]: I1002 09:48:15.250389 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:48:15 crc kubenswrapper[5035]: I1002 09:48:15.888765 5035 generic.go:334] "Generic (PLEG): container finished" podID="0307b78d-517d-4442-a0d7-1efa3f46098f" containerID="7665b703df59e2a68f83503c6bbe8a03e984c54b6e9540d89ec79d2dde08de0a" exitCode=0 Oct 02 09:48:15 crc kubenswrapper[5035]: I1002 09:48:15.888803 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerDied","Data":"7665b703df59e2a68f83503c6bbe8a03e984c54b6e9540d89ec79d2dde08de0a"} Oct 02 09:48:16 crc kubenswrapper[5035]: I1002 09:48:16.923196 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerStarted","Data":"694b7d02fc7eaf1eb45a182928ec76d61565c77cde565ce5a6ae7c329f67dcd0"} Oct 02 09:48:17 crc kubenswrapper[5035]: I1002 09:48:17.951439 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerStarted","Data":"069e6d39a78483b6b54a47736457d3ca3d16689f6a2da602802765c8dcd851e7"} Oct 02 09:48:17 crc kubenswrapper[5035]: I1002 09:48:17.952409 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"0307b78d-517d-4442-a0d7-1efa3f46098f","Type":"ContainerStarted","Data":"ee09137e8f72b3d7546ea06f104afd8de725835e21240a6b68464f400c35ec72"} Oct 02 09:48:17 crc kubenswrapper[5035]: I1002 09:48:17.955778 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Oct 02 09:48:17 crc kubenswrapper[5035]: I1002 09:48:17.955833 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Oct 02 09:48:18 crc kubenswrapper[5035]: I1002 09:48:18.006471 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-conductor-0" podStartSLOduration=108.942312675 podStartE2EDuration="2m29.006451547s" podCreationTimestamp="2025-10-02 09:45:49 +0000 UTC" firstStartedPulling="2025-10-02 09:45:54.235750407 +0000 UTC m=+1119.592094432" lastFinishedPulling="2025-10-02 09:46:34.299889279 +0000 UTC m=+1159.656233304" observedRunningTime="2025-10-02 09:48:18.006343483 +0000 UTC m=+1263.362687518" watchObservedRunningTime="2025-10-02 09:48:18.006451547 +0000 UTC m=+1263.362795582" Oct 02 09:48:18 crc kubenswrapper[5035]: I1002 09:48:18.383872 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 09:48:18 crc kubenswrapper[5035]: I1002 09:48:18.388393 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 09:48:18 crc kubenswrapper[5035]: I1002 09:48:18.391527 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 09:48:18 crc kubenswrapper[5035]: I1002 09:48:18.968405 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.924568 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.974459 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-combined-ca-bundle\") pod \"9a986167-8d84-4b52-91a1-5858e82b734f\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.974525 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n28mm\" (UniqueName: \"kubernetes.io/projected/9a986167-8d84-4b52-91a1-5858e82b734f-kube-api-access-n28mm\") pod \"9a986167-8d84-4b52-91a1-5858e82b734f\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.974841 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-config-data\") pod \"9a986167-8d84-4b52-91a1-5858e82b734f\" (UID: \"9a986167-8d84-4b52-91a1-5858e82b734f\") " Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.981864 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a986167-8d84-4b52-91a1-5858e82b734f-kube-api-access-n28mm" (OuterVolumeSpecName: "kube-api-access-n28mm") pod "9a986167-8d84-4b52-91a1-5858e82b734f" (UID: "9a986167-8d84-4b52-91a1-5858e82b734f"). InnerVolumeSpecName "kube-api-access-n28mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.985200 5035 generic.go:334] "Generic (PLEG): container finished" podID="9a986167-8d84-4b52-91a1-5858e82b734f" containerID="b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc" exitCode=137 Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.985266 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.985343 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9a986167-8d84-4b52-91a1-5858e82b734f","Type":"ContainerDied","Data":"b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc"} Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.985447 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9a986167-8d84-4b52-91a1-5858e82b734f","Type":"ContainerDied","Data":"c56c76315c0301a5b3f21073a4f357a887a5099092be29e53c8d68f212c7e77f"} Oct 02 09:48:20 crc kubenswrapper[5035]: I1002 09:48:20.985475 5035 scope.go:117] "RemoveContainer" containerID="b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.012482 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a986167-8d84-4b52-91a1-5858e82b734f" (UID: "9a986167-8d84-4b52-91a1-5858e82b734f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.018313 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-config-data" (OuterVolumeSpecName: "config-data") pod "9a986167-8d84-4b52-91a1-5858e82b734f" (UID: "9a986167-8d84-4b52-91a1-5858e82b734f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.059673 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.077632 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.077664 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a986167-8d84-4b52-91a1-5858e82b734f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.077674 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n28mm\" (UniqueName: \"kubernetes.io/projected/9a986167-8d84-4b52-91a1-5858e82b734f-kube-api-access-n28mm\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.114258 5035 scope.go:117] "RemoveContainer" containerID="b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc" Oct 02 09:48:21 crc kubenswrapper[5035]: E1002 09:48:21.114675 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc\": container with ID starting with b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc not found: ID does not exist" containerID="b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.114724 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc"} err="failed to get container status \"b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc\": rpc error: code = NotFound desc = could not find container \"b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc\": container with ID starting with b287e8d4eff9e14ee804a495325e0e2c54b09cd5f2331ebadf3e057c060438fc not found: ID does not exist" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.316215 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.325120 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.336192 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:48:21 crc kubenswrapper[5035]: E1002 09:48:21.336733 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a986167-8d84-4b52-91a1-5858e82b734f" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.336760 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a986167-8d84-4b52-91a1-5858e82b734f" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.337009 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a986167-8d84-4b52-91a1-5858e82b734f" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.337837 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.339987 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.340124 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.343340 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.348240 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.383982 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.384041 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.384140 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5rsq\" (UniqueName: \"kubernetes.io/projected/e316d9ab-3baf-429e-9839-e7dea3125ca4-kube-api-access-g5rsq\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.384191 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.384281 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.485676 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.485760 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.485797 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.485844 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5rsq\" (UniqueName: \"kubernetes.io/projected/e316d9ab-3baf-429e-9839-e7dea3125ca4-kube-api-access-g5rsq\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.485874 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.490128 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.490253 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.490404 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.490794 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e316d9ab-3baf-429e-9839-e7dea3125ca4-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.502808 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5rsq\" (UniqueName: \"kubernetes.io/projected/e316d9ab-3baf-429e-9839-e7dea3125ca4-kube-api-access-g5rsq\") pod \"nova-cell1-novncproxy-0\" (UID: \"e316d9ab-3baf-429e-9839-e7dea3125ca4\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:21 crc kubenswrapper[5035]: I1002 09:48:21.653144 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:22 crc kubenswrapper[5035]: I1002 09:48:22.075155 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 09:48:22 crc kubenswrapper[5035]: I1002 09:48:22.188132 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a986167-8d84-4b52-91a1-5858e82b734f" path="/var/lib/kubelet/pods/9a986167-8d84-4b52-91a1-5858e82b734f/volumes" Oct 02 09:48:23 crc kubenswrapper[5035]: I1002 09:48:23.006494 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e316d9ab-3baf-429e-9839-e7dea3125ca4","Type":"ContainerStarted","Data":"fc50e0b2bb426044f2bba60a8826836e845eb07bec8668059438526c201c78ee"} Oct 02 09:48:23 crc kubenswrapper[5035]: I1002 09:48:23.006849 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e316d9ab-3baf-429e-9839-e7dea3125ca4","Type":"ContainerStarted","Data":"fd6226bfc81b34f9714f9938bf7300a2cf1d0488db21ce6eafc95c4e1e0a9e9c"} Oct 02 09:48:23 crc kubenswrapper[5035]: I1002 09:48:23.038654 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.038631285 podStartE2EDuration="2.038631285s" podCreationTimestamp="2025-10-02 09:48:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:23.022957099 +0000 UTC m=+1268.379301134" watchObservedRunningTime="2025-10-02 09:48:23.038631285 +0000 UTC m=+1268.394975310" Oct 02 09:48:24 crc kubenswrapper[5035]: I1002 09:48:24.173470 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 09:48:24 crc kubenswrapper[5035]: I1002 09:48:24.174099 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 09:48:24 crc kubenswrapper[5035]: I1002 09:48:24.175332 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 09:48:24 crc kubenswrapper[5035]: I1002 09:48:24.177093 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.026828 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.030171 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.225553 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-lmk4b"] Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.227501 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.251685 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-lmk4b"] Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.364223 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wnmx\" (UniqueName: \"kubernetes.io/projected/85b5680e-bfe0-4c34-b773-14364d25cac5-kube-api-access-7wnmx\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.364520 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-config\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.364595 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.364744 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.364809 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.364878 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.466812 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.466865 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.466907 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.466975 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wnmx\" (UniqueName: \"kubernetes.io/projected/85b5680e-bfe0-4c34-b773-14364d25cac5-kube-api-access-7wnmx\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.467052 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-config\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.467068 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.467907 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.468318 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.468315 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.468315 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.468435 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85b5680e-bfe0-4c34-b773-14364d25cac5-config\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.488303 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wnmx\" (UniqueName: \"kubernetes.io/projected/85b5680e-bfe0-4c34-b773-14364d25cac5-kube-api-access-7wnmx\") pod \"dnsmasq-dns-59cf4bdb65-lmk4b\" (UID: \"85b5680e-bfe0-4c34-b773-14364d25cac5\") " pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:25 crc kubenswrapper[5035]: I1002 09:48:25.569405 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:26 crc kubenswrapper[5035]: I1002 09:48:26.058449 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-lmk4b"] Oct 02 09:48:26 crc kubenswrapper[5035]: W1002 09:48:26.062979 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85b5680e_bfe0_4c34_b773_14364d25cac5.slice/crio-b0613c40a6789f632015df2f2c255b1b02be5d133cc31f697f4405990ad203e5 WatchSource:0}: Error finding container b0613c40a6789f632015df2f2c255b1b02be5d133cc31f697f4405990ad203e5: Status 404 returned error can't find the container with id b0613c40a6789f632015df2f2c255b1b02be5d133cc31f697f4405990ad203e5 Oct 02 09:48:26 crc kubenswrapper[5035]: I1002 09:48:26.653937 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.048222 5035 generic.go:334] "Generic (PLEG): container finished" podID="85b5680e-bfe0-4c34-b773-14364d25cac5" containerID="698f41cd11553ec9936a7dadbba1ff4a7c7b3e84a499a3adf24616b43597b064" exitCode=0 Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.049686 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" event={"ID":"85b5680e-bfe0-4c34-b773-14364d25cac5","Type":"ContainerDied","Data":"698f41cd11553ec9936a7dadbba1ff4a7c7b3e84a499a3adf24616b43597b064"} Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.049763 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" event={"ID":"85b5680e-bfe0-4c34-b773-14364d25cac5","Type":"ContainerStarted","Data":"b0613c40a6789f632015df2f2c255b1b02be5d133cc31f697f4405990ad203e5"} Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.618133 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.618652 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-central-agent" containerID="cri-o://7a0d7223cb9ab44c4050171021800a688f29f1adc2a5d23300b979ae7da53649" gracePeriod=30 Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.618765 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="sg-core" containerID="cri-o://25cfb55874b0f5f5155071b0d0f5babe50bb2c756a12cdaad26e799c30b34675" gracePeriod=30 Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.618863 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="proxy-httpd" containerID="cri-o://0f5c7f8199a7e7defdc74a072725c05bfe2e2b6fb5c72fbb2f025ad1d24fc416" gracePeriod=30 Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.618950 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-notification-agent" containerID="cri-o://0768287e03f6c2b26503dc517804a61b5f0dac3818c4593be9246059e8f3a7fa" gracePeriod=30 Oct 02 09:48:27 crc kubenswrapper[5035]: I1002 09:48:27.814755 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.060514 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" event={"ID":"85b5680e-bfe0-4c34-b773-14364d25cac5","Type":"ContainerStarted","Data":"91e954645e15d958b97dfbc3410f41488dbb73c2ad002a2a3cfeecef80c563a7"} Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.061588 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071229 5035 generic.go:334] "Generic (PLEG): container finished" podID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerID="0f5c7f8199a7e7defdc74a072725c05bfe2e2b6fb5c72fbb2f025ad1d24fc416" exitCode=0 Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071261 5035 generic.go:334] "Generic (PLEG): container finished" podID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerID="25cfb55874b0f5f5155071b0d0f5babe50bb2c756a12cdaad26e799c30b34675" exitCode=2 Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071271 5035 generic.go:334] "Generic (PLEG): container finished" podID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerID="0768287e03f6c2b26503dc517804a61b5f0dac3818c4593be9246059e8f3a7fa" exitCode=0 Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071279 5035 generic.go:334] "Generic (PLEG): container finished" podID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerID="7a0d7223cb9ab44c4050171021800a688f29f1adc2a5d23300b979ae7da53649" exitCode=0 Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071444 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-log" containerID="cri-o://0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc" gracePeriod=30 Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071597 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerDied","Data":"0f5c7f8199a7e7defdc74a072725c05bfe2e2b6fb5c72fbb2f025ad1d24fc416"} Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071634 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerDied","Data":"25cfb55874b0f5f5155071b0d0f5babe50bb2c756a12cdaad26e799c30b34675"} Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071648 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerDied","Data":"0768287e03f6c2b26503dc517804a61b5f0dac3818c4593be9246059e8f3a7fa"} Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071661 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerDied","Data":"7a0d7223cb9ab44c4050171021800a688f29f1adc2a5d23300b979ae7da53649"} Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.071632 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-api" containerID="cri-o://6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76" gracePeriod=30 Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.089436 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" podStartSLOduration=3.089415027 podStartE2EDuration="3.089415027s" podCreationTimestamp="2025-10-02 09:48:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:28.08060472 +0000 UTC m=+1273.436948745" watchObservedRunningTime="2025-10-02 09:48:28.089415027 +0000 UTC m=+1273.445759052" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.413630 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.544265 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hblkg\" (UniqueName: \"kubernetes.io/projected/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-kube-api-access-hblkg\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.544381 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-ceilometer-tls-certs\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.544417 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-config-data\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.545205 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-sg-core-conf-yaml\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.545306 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-log-httpd\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.545418 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-combined-ca-bundle\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.545457 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-run-httpd\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.545494 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-scripts\") pod \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\" (UID: \"ccd06cdd-aba8-4985-aa4c-499bbd9b118e\") " Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.546451 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.546676 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.551217 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-scripts" (OuterVolumeSpecName: "scripts") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.551782 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-kube-api-access-hblkg" (OuterVolumeSpecName: "kube-api-access-hblkg") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "kube-api-access-hblkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.612081 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.625434 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.650204 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.650243 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.650255 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hblkg\" (UniqueName: \"kubernetes.io/projected/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-kube-api-access-hblkg\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.650268 5035 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.650279 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.650302 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.651981 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.688429 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-config-data" (OuterVolumeSpecName: "config-data") pod "ccd06cdd-aba8-4985-aa4c-499bbd9b118e" (UID: "ccd06cdd-aba8-4985-aa4c-499bbd9b118e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.751788 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:28 crc kubenswrapper[5035]: I1002 09:48:28.752114 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccd06cdd-aba8-4985-aa4c-499bbd9b118e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.083780 5035 generic.go:334] "Generic (PLEG): container finished" podID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerID="0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc" exitCode=143 Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.083867 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27578bd8-1ea0-4dd8-bef6-c1cd69815419","Type":"ContainerDied","Data":"0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc"} Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.086786 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ccd06cdd-aba8-4985-aa4c-499bbd9b118e","Type":"ContainerDied","Data":"3887d0dfafbf976500675857b452965300326f20b7255bb676db9b6a329f62af"} Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.086825 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.086838 5035 scope.go:117] "RemoveContainer" containerID="0f5c7f8199a7e7defdc74a072725c05bfe2e2b6fb5c72fbb2f025ad1d24fc416" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.112353 5035 scope.go:117] "RemoveContainer" containerID="25cfb55874b0f5f5155071b0d0f5babe50bb2c756a12cdaad26e799c30b34675" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.150553 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.176988 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.197151 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:29 crc kubenswrapper[5035]: E1002 09:48:29.197672 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-notification-agent" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.197733 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-notification-agent" Oct 02 09:48:29 crc kubenswrapper[5035]: E1002 09:48:29.197767 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-central-agent" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.197776 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-central-agent" Oct 02 09:48:29 crc kubenswrapper[5035]: E1002 09:48:29.197791 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="sg-core" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.197797 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="sg-core" Oct 02 09:48:29 crc kubenswrapper[5035]: E1002 09:48:29.197812 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="proxy-httpd" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.197818 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="proxy-httpd" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.198000 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-notification-agent" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.198022 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="proxy-httpd" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.198034 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="ceilometer-central-agent" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.198053 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" containerName="sg-core" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.199763 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.204496 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.204710 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.204828 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.204908 5035 scope.go:117] "RemoveContainer" containerID="0768287e03f6c2b26503dc517804a61b5f0dac3818c4593be9246059e8f3a7fa" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.220897 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.231749 5035 scope.go:117] "RemoveContainer" containerID="7a0d7223cb9ab44c4050171021800a688f29f1adc2a5d23300b979ae7da53649" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.262999 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-run-httpd\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.263184 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-log-httpd\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.263302 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-scripts\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.263417 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-config-data\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.263492 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.263577 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47grw\" (UniqueName: \"kubernetes.io/projected/9ecf357b-db63-4382-a77c-562a2f48cecb-kube-api-access-47grw\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.263649 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.263678 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.365147 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47grw\" (UniqueName: \"kubernetes.io/projected/9ecf357b-db63-4382-a77c-562a2f48cecb-kube-api-access-47grw\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.365385 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.365485 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.365618 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-run-httpd\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.365759 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-log-httpd\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.365902 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-scripts\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.366020 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-config-data\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.366133 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.366067 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-run-httpd\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.366290 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-log-httpd\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.369941 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-scripts\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.370390 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.370621 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.371384 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.371412 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-config-data\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.391447 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47grw\" (UniqueName: \"kubernetes.io/projected/9ecf357b-db63-4382-a77c-562a2f48cecb-kube-api-access-47grw\") pod \"ceilometer-0\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.518857 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.767912 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:29 crc kubenswrapper[5035]: W1002 09:48:29.966836 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ecf357b_db63_4382_a77c_562a2f48cecb.slice/crio-25ed9c5b086381508ddb099caa0af3cb539448abdafc63274b1f2b0e57b5bc62 WatchSource:0}: Error finding container 25ed9c5b086381508ddb099caa0af3cb539448abdafc63274b1f2b0e57b5bc62: Status 404 returned error can't find the container with id 25ed9c5b086381508ddb099caa0af3cb539448abdafc63274b1f2b0e57b5bc62 Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.969313 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:29 crc kubenswrapper[5035]: I1002 09:48:29.970298 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:48:30 crc kubenswrapper[5035]: I1002 09:48:30.098340 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerStarted","Data":"25ed9c5b086381508ddb099caa0af3cb539448abdafc63274b1f2b0e57b5bc62"} Oct 02 09:48:30 crc kubenswrapper[5035]: I1002 09:48:30.181153 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccd06cdd-aba8-4985-aa4c-499bbd9b118e" path="/var/lib/kubelet/pods/ccd06cdd-aba8-4985-aa4c-499bbd9b118e/volumes" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.110787 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerStarted","Data":"c09ce0b256e5d90e50314bc158d09f0eddede663e178107504e385cc41220f52"} Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.653905 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.682313 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.721013 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.909743 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-config-data\") pod \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.910110 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdd8m\" (UniqueName: \"kubernetes.io/projected/27578bd8-1ea0-4dd8-bef6-c1cd69815419-kube-api-access-mdd8m\") pod \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.910180 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27578bd8-1ea0-4dd8-bef6-c1cd69815419-logs\") pod \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.910232 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-combined-ca-bundle\") pod \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\" (UID: \"27578bd8-1ea0-4dd8-bef6-c1cd69815419\") " Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.910589 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27578bd8-1ea0-4dd8-bef6-c1cd69815419-logs" (OuterVolumeSpecName: "logs") pod "27578bd8-1ea0-4dd8-bef6-c1cd69815419" (UID: "27578bd8-1ea0-4dd8-bef6-c1cd69815419"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.910910 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27578bd8-1ea0-4dd8-bef6-c1cd69815419-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.915712 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27578bd8-1ea0-4dd8-bef6-c1cd69815419-kube-api-access-mdd8m" (OuterVolumeSpecName: "kube-api-access-mdd8m") pod "27578bd8-1ea0-4dd8-bef6-c1cd69815419" (UID: "27578bd8-1ea0-4dd8-bef6-c1cd69815419"). InnerVolumeSpecName "kube-api-access-mdd8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.950015 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27578bd8-1ea0-4dd8-bef6-c1cd69815419" (UID: "27578bd8-1ea0-4dd8-bef6-c1cd69815419"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:31 crc kubenswrapper[5035]: I1002 09:48:31.958754 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-config-data" (OuterVolumeSpecName: "config-data") pod "27578bd8-1ea0-4dd8-bef6-c1cd69815419" (UID: "27578bd8-1ea0-4dd8-bef6-c1cd69815419"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.012550 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.012593 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27578bd8-1ea0-4dd8-bef6-c1cd69815419-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.012605 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdd8m\" (UniqueName: \"kubernetes.io/projected/27578bd8-1ea0-4dd8-bef6-c1cd69815419-kube-api-access-mdd8m\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.123953 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerStarted","Data":"35746bdb68bb8babda9b723a6b9201cd3834f755fb27faa1404263889d05b2e0"} Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.134095 5035 generic.go:334] "Generic (PLEG): container finished" podID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerID="6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76" exitCode=0 Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.135634 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.136386 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27578bd8-1ea0-4dd8-bef6-c1cd69815419","Type":"ContainerDied","Data":"6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76"} Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.136425 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"27578bd8-1ea0-4dd8-bef6-c1cd69815419","Type":"ContainerDied","Data":"aa9f98e434c3f7df8f9e65b9f2e620007aa00da43e5135d8698073bbbd8d156d"} Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.136449 5035 scope.go:117] "RemoveContainer" containerID="6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.151096 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.157768 5035 scope.go:117] "RemoveContainer" containerID="0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.194365 5035 scope.go:117] "RemoveContainer" containerID="6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76" Oct 02 09:48:32 crc kubenswrapper[5035]: E1002 09:48:32.194877 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76\": container with ID starting with 6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76 not found: ID does not exist" containerID="6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.194911 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76"} err="failed to get container status \"6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76\": rpc error: code = NotFound desc = could not find container \"6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76\": container with ID starting with 6de80f74bfe6a293724e887d97f749bb0503952025eea9cd74722e1ae4d25b76 not found: ID does not exist" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.194931 5035 scope.go:117] "RemoveContainer" containerID="0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc" Oct 02 09:48:32 crc kubenswrapper[5035]: E1002 09:48:32.195340 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc\": container with ID starting with 0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc not found: ID does not exist" containerID="0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.195370 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc"} err="failed to get container status \"0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc\": rpc error: code = NotFound desc = could not find container \"0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc\": container with ID starting with 0926b4517370b94d3e7457685e3c033e6e2be2c9916ad31abfd385d664949bdc not found: ID does not exist" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.244773 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.285692 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.316685 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:32 crc kubenswrapper[5035]: E1002 09:48:32.337268 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-api" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.337470 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-api" Oct 02 09:48:32 crc kubenswrapper[5035]: E1002 09:48:32.337512 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-log" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.337522 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-log" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.338030 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-log" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.338079 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" containerName="nova-api-api" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.340070 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.350818 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.360738 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.366035 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.367439 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.379884 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dsl64"] Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.382432 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.386201 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.387415 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.397237 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsl64"] Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.528888 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-config-data\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.529439 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1ebd571-0a0d-473e-b502-b3727de8b617-logs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.529485 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5k5m\" (UniqueName: \"kubernetes.io/projected/093232c1-8151-4eac-97ec-76c29f228e05-kube-api-access-w5k5m\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.529915 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgl74\" (UniqueName: \"kubernetes.io/projected/c1ebd571-0a0d-473e-b502-b3727de8b617-kube-api-access-rgl74\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.530004 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.530098 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.530194 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.530472 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-config-data\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.530550 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-scripts\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.530702 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.632769 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-config-data\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.632830 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1ebd571-0a0d-473e-b502-b3727de8b617-logs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.632857 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5k5m\" (UniqueName: \"kubernetes.io/projected/093232c1-8151-4eac-97ec-76c29f228e05-kube-api-access-w5k5m\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.632886 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgl74\" (UniqueName: \"kubernetes.io/projected/c1ebd571-0a0d-473e-b502-b3727de8b617-kube-api-access-rgl74\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.632930 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.632979 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.632996 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.633039 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-config-data\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.633062 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-scripts\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.633101 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.633996 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1ebd571-0a0d-473e-b502-b3727de8b617-logs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.638868 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.639919 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-config-data\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.640547 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-public-tls-certs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.640561 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-config-data\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.641701 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-scripts\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.644090 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.649565 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.651711 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgl74\" (UniqueName: \"kubernetes.io/projected/c1ebd571-0a0d-473e-b502-b3727de8b617-kube-api-access-rgl74\") pod \"nova-api-0\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.655598 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5k5m\" (UniqueName: \"kubernetes.io/projected/093232c1-8151-4eac-97ec-76c29f228e05-kube-api-access-w5k5m\") pod \"nova-cell1-cell-mapping-dsl64\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.679105 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:32 crc kubenswrapper[5035]: I1002 09:48:32.704726 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:33 crc kubenswrapper[5035]: I1002 09:48:33.149901 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerStarted","Data":"5b5131abcbb5bc8ef131d86e4a1556e66b7d39654c483641b0dfce473cb447dc"} Oct 02 09:48:33 crc kubenswrapper[5035]: I1002 09:48:33.216813 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:33 crc kubenswrapper[5035]: W1002 09:48:33.219018 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1ebd571_0a0d_473e_b502_b3727de8b617.slice/crio-2ca4bd645d25ee623ebcf93f265799ea78a41068dbeac583cf9ba0e7d92cfdae WatchSource:0}: Error finding container 2ca4bd645d25ee623ebcf93f265799ea78a41068dbeac583cf9ba0e7d92cfdae: Status 404 returned error can't find the container with id 2ca4bd645d25ee623ebcf93f265799ea78a41068dbeac583cf9ba0e7d92cfdae Oct 02 09:48:33 crc kubenswrapper[5035]: I1002 09:48:33.297513 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsl64"] Oct 02 09:48:33 crc kubenswrapper[5035]: W1002 09:48:33.300225 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod093232c1_8151_4eac_97ec_76c29f228e05.slice/crio-31c312094181789ed3f53ec924cf351ebb8d3f75444de441784d012a70f1ecd4 WatchSource:0}: Error finding container 31c312094181789ed3f53ec924cf351ebb8d3f75444de441784d012a70f1ecd4: Status 404 returned error can't find the container with id 31c312094181789ed3f53ec924cf351ebb8d3f75444de441784d012a70f1ecd4 Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.161440 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsl64" event={"ID":"093232c1-8151-4eac-97ec-76c29f228e05","Type":"ContainerStarted","Data":"712e3a0139796a4a1434856e06363d8133baf9310624eb20d5916f30f3d4fd86"} Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.161840 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsl64" event={"ID":"093232c1-8151-4eac-97ec-76c29f228e05","Type":"ContainerStarted","Data":"31c312094181789ed3f53ec924cf351ebb8d3f75444de441784d012a70f1ecd4"} Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.170468 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-central-agent" containerID="cri-o://c09ce0b256e5d90e50314bc158d09f0eddede663e178107504e385cc41220f52" gracePeriod=30 Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.170509 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-notification-agent" containerID="cri-o://35746bdb68bb8babda9b723a6b9201cd3834f755fb27faa1404263889d05b2e0" gracePeriod=30 Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.170507 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="proxy-httpd" containerID="cri-o://5383078a22cfae40ed3d838e37ee6a049538d61cfb7b6391bfb86beaa7240088" gracePeriod=30 Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.170506 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="sg-core" containerID="cri-o://5b5131abcbb5bc8ef131d86e4a1556e66b7d39654c483641b0dfce473cb447dc" gracePeriod=30 Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.186671 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27578bd8-1ea0-4dd8-bef6-c1cd69815419" path="/var/lib/kubelet/pods/27578bd8-1ea0-4dd8-bef6-c1cd69815419/volumes" Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.187349 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.187420 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1ebd571-0a0d-473e-b502-b3727de8b617","Type":"ContainerStarted","Data":"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803"} Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.187515 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1ebd571-0a0d-473e-b502-b3727de8b617","Type":"ContainerStarted","Data":"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb"} Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.187639 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1ebd571-0a0d-473e-b502-b3727de8b617","Type":"ContainerStarted","Data":"2ca4bd645d25ee623ebcf93f265799ea78a41068dbeac583cf9ba0e7d92cfdae"} Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.187692 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerStarted","Data":"5383078a22cfae40ed3d838e37ee6a049538d61cfb7b6391bfb86beaa7240088"} Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.203662 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dsl64" podStartSLOduration=2.203521612 podStartE2EDuration="2.203521612s" podCreationTimestamp="2025-10-02 09:48:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:34.186931498 +0000 UTC m=+1279.543275523" watchObservedRunningTime="2025-10-02 09:48:34.203521612 +0000 UTC m=+1279.559865637" Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.221460 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.221439944 podStartE2EDuration="2.221439944s" podCreationTimestamp="2025-10-02 09:48:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:34.204330455 +0000 UTC m=+1279.560674480" watchObservedRunningTime="2025-10-02 09:48:34.221439944 +0000 UTC m=+1279.577783969" Oct 02 09:48:34 crc kubenswrapper[5035]: I1002 09:48:34.251715 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.705601858 podStartE2EDuration="5.251692295s" podCreationTimestamp="2025-10-02 09:48:29 +0000 UTC" firstStartedPulling="2025-10-02 09:48:29.969996084 +0000 UTC m=+1275.326340109" lastFinishedPulling="2025-10-02 09:48:33.516086521 +0000 UTC m=+1278.872430546" observedRunningTime="2025-10-02 09:48:34.225040609 +0000 UTC m=+1279.581384634" watchObservedRunningTime="2025-10-02 09:48:34.251692295 +0000 UTC m=+1279.608036310" Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.187305 5035 generic.go:334] "Generic (PLEG): container finished" podID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerID="5383078a22cfae40ed3d838e37ee6a049538d61cfb7b6391bfb86beaa7240088" exitCode=0 Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.187752 5035 generic.go:334] "Generic (PLEG): container finished" podID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerID="5b5131abcbb5bc8ef131d86e4a1556e66b7d39654c483641b0dfce473cb447dc" exitCode=2 Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.187771 5035 generic.go:334] "Generic (PLEG): container finished" podID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerID="35746bdb68bb8babda9b723a6b9201cd3834f755fb27faa1404263889d05b2e0" exitCode=0 Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.189182 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerDied","Data":"5383078a22cfae40ed3d838e37ee6a049538d61cfb7b6391bfb86beaa7240088"} Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.189234 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerDied","Data":"5b5131abcbb5bc8ef131d86e4a1556e66b7d39654c483641b0dfce473cb447dc"} Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.189257 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerDied","Data":"35746bdb68bb8babda9b723a6b9201cd3834f755fb27faa1404263889d05b2e0"} Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.571810 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-lmk4b" Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.653596 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-759js"] Oct 02 09:48:35 crc kubenswrapper[5035]: I1002 09:48:35.653878 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-759js" podUID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerName="dnsmasq-dns" containerID="cri-o://2b3daca74aad2f3efdfff4772d258e1591628504c5da8ea5ab47a8bf55fa16aa" gracePeriod=10 Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.200234 5035 generic.go:334] "Generic (PLEG): container finished" podID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerID="2b3daca74aad2f3efdfff4772d258e1591628504c5da8ea5ab47a8bf55fa16aa" exitCode=0 Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.200790 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-759js" event={"ID":"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c","Type":"ContainerDied","Data":"2b3daca74aad2f3efdfff4772d258e1591628504c5da8ea5ab47a8bf55fa16aa"} Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.200819 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-759js" event={"ID":"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c","Type":"ContainerDied","Data":"68ec136eee57ed7f57da5104f7c3cea8bd009915055a0d1a4de36f8a780e48d0"} Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.200831 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68ec136eee57ed7f57da5104f7c3cea8bd009915055a0d1a4de36f8a780e48d0" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.204781 5035 generic.go:334] "Generic (PLEG): container finished" podID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerID="c09ce0b256e5d90e50314bc158d09f0eddede663e178107504e385cc41220f52" exitCode=0 Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.205665 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerDied","Data":"c09ce0b256e5d90e50314bc158d09f0eddede663e178107504e385cc41220f52"} Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.205734 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ecf357b-db63-4382-a77c-562a2f48cecb","Type":"ContainerDied","Data":"25ed9c5b086381508ddb099caa0af3cb539448abdafc63274b1f2b0e57b5bc62"} Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.205750 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25ed9c5b086381508ddb099caa0af3cb539448abdafc63274b1f2b0e57b5bc62" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.279014 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.283878 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.419925 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47grw\" (UniqueName: \"kubernetes.io/projected/9ecf357b-db63-4382-a77c-562a2f48cecb-kube-api-access-47grw\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420057 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-sb\") pod \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420096 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-config-data\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420134 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7swjd\" (UniqueName: \"kubernetes.io/projected/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-kube-api-access-7swjd\") pod \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420178 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-sg-core-conf-yaml\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420204 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-nb\") pod \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420259 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-config\") pod \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420307 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-swift-storage-0\") pod \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420339 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-ceilometer-tls-certs\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420384 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-svc\") pod \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\" (UID: \"5c26a283-a4b2-433a-9d94-4d5cddfc5e9c\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420407 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-log-httpd\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420432 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-combined-ca-bundle\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420475 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-scripts\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.420507 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-run-httpd\") pod \"9ecf357b-db63-4382-a77c-562a2f48cecb\" (UID: \"9ecf357b-db63-4382-a77c-562a2f48cecb\") " Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.421500 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.421877 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.426435 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ecf357b-db63-4382-a77c-562a2f48cecb-kube-api-access-47grw" (OuterVolumeSpecName: "kube-api-access-47grw") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "kube-api-access-47grw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.438436 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-kube-api-access-7swjd" (OuterVolumeSpecName: "kube-api-access-7swjd") pod "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" (UID: "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c"). InnerVolumeSpecName "kube-api-access-7swjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.452119 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-scripts" (OuterVolumeSpecName: "scripts") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.498009 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.522325 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7swjd\" (UniqueName: \"kubernetes.io/projected/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-kube-api-access-7swjd\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.522353 5035 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.522362 5035 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.522370 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.522378 5035 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ecf357b-db63-4382-a77c-562a2f48cecb-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.522387 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47grw\" (UniqueName: \"kubernetes.io/projected/9ecf357b-db63-4382-a77c-562a2f48cecb-kube-api-access-47grw\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.529647 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.533999 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-config" (OuterVolumeSpecName: "config") pod "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" (UID: "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.536195 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" (UID: "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.539807 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" (UID: "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.543645 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" (UID: "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.566330 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" (UID: "5c26a283-a4b2-433a-9d94-4d5cddfc5e9c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.566604 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.596124 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-config-data" (OuterVolumeSpecName: "config-data") pod "9ecf357b-db63-4382-a77c-562a2f48cecb" (UID: "9ecf357b-db63-4382-a77c-562a2f48cecb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624013 5035 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624052 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624065 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624076 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624090 5035 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624101 5035 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-config\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624113 5035 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:36 crc kubenswrapper[5035]: I1002 09:48:36.624124 5035 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ecf357b-db63-4382-a77c-562a2f48cecb-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.213406 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.213440 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-759js" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.252928 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-759js"] Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.266277 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-759js"] Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.274181 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.289031 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.300180 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:37 crc kubenswrapper[5035]: E1002 09:48:37.300676 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-notification-agent" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.300697 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-notification-agent" Oct 02 09:48:37 crc kubenswrapper[5035]: E1002 09:48:37.300732 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerName="init" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.300740 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerName="init" Oct 02 09:48:37 crc kubenswrapper[5035]: E1002 09:48:37.300757 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-central-agent" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.300764 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-central-agent" Oct 02 09:48:37 crc kubenswrapper[5035]: E1002 09:48:37.300785 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="proxy-httpd" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.300791 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="proxy-httpd" Oct 02 09:48:37 crc kubenswrapper[5035]: E1002 09:48:37.300804 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="sg-core" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.300811 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="sg-core" Oct 02 09:48:37 crc kubenswrapper[5035]: E1002 09:48:37.300831 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerName="dnsmasq-dns" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.300839 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerName="dnsmasq-dns" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.301037 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="sg-core" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.301053 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="proxy-httpd" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.301072 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-central-agent" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.301093 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" containerName="dnsmasq-dns" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.301101 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" containerName="ceilometer-notification-agent" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.303550 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.315093 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.315294 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.315309 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.315334 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441143 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-config-data\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441476 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0400439-a652-4472-9f8a-51323ea10780-run-httpd\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441503 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441663 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441683 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-scripts\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441698 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441778 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47l7n\" (UniqueName: \"kubernetes.io/projected/a0400439-a652-4472-9f8a-51323ea10780-kube-api-access-47l7n\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.441800 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0400439-a652-4472-9f8a-51323ea10780-log-httpd\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.544817 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0400439-a652-4472-9f8a-51323ea10780-run-httpd\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.544907 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.545057 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.545096 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-scripts\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.545123 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.545173 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47l7n\" (UniqueName: \"kubernetes.io/projected/a0400439-a652-4472-9f8a-51323ea10780-kube-api-access-47l7n\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.545200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0400439-a652-4472-9f8a-51323ea10780-log-httpd\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.545258 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-config-data\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.545504 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0400439-a652-4472-9f8a-51323ea10780-run-httpd\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.546356 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a0400439-a652-4472-9f8a-51323ea10780-log-httpd\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.552641 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-scripts\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.553358 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-config-data\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.563053 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.567373 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.568034 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0400439-a652-4472-9f8a-51323ea10780-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.570562 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47l7n\" (UniqueName: \"kubernetes.io/projected/a0400439-a652-4472-9f8a-51323ea10780-kube-api-access-47l7n\") pod \"ceilometer-0\" (UID: \"a0400439-a652-4472-9f8a-51323ea10780\") " pod="openstack/ceilometer-0" Oct 02 09:48:37 crc kubenswrapper[5035]: I1002 09:48:37.626977 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 09:48:38 crc kubenswrapper[5035]: I1002 09:48:38.089752 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 09:48:38 crc kubenswrapper[5035]: W1002 09:48:38.095831 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0400439_a652_4472_9f8a_51323ea10780.slice/crio-11498d4ea8f80fabeac957e32db5f0b2c2a6f2638235ee1a7ef138228891ac52 WatchSource:0}: Error finding container 11498d4ea8f80fabeac957e32db5f0b2c2a6f2638235ee1a7ef138228891ac52: Status 404 returned error can't find the container with id 11498d4ea8f80fabeac957e32db5f0b2c2a6f2638235ee1a7ef138228891ac52 Oct 02 09:48:38 crc kubenswrapper[5035]: I1002 09:48:38.181956 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c26a283-a4b2-433a-9d94-4d5cddfc5e9c" path="/var/lib/kubelet/pods/5c26a283-a4b2-433a-9d94-4d5cddfc5e9c/volumes" Oct 02 09:48:38 crc kubenswrapper[5035]: I1002 09:48:38.183312 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ecf357b-db63-4382-a77c-562a2f48cecb" path="/var/lib/kubelet/pods/9ecf357b-db63-4382-a77c-562a2f48cecb/volumes" Oct 02 09:48:38 crc kubenswrapper[5035]: I1002 09:48:38.226151 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0400439-a652-4472-9f8a-51323ea10780","Type":"ContainerStarted","Data":"11498d4ea8f80fabeac957e32db5f0b2c2a6f2638235ee1a7ef138228891ac52"} Oct 02 09:48:39 crc kubenswrapper[5035]: I1002 09:48:39.242793 5035 generic.go:334] "Generic (PLEG): container finished" podID="093232c1-8151-4eac-97ec-76c29f228e05" containerID="712e3a0139796a4a1434856e06363d8133baf9310624eb20d5916f30f3d4fd86" exitCode=0 Oct 02 09:48:39 crc kubenswrapper[5035]: I1002 09:48:39.242903 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsl64" event={"ID":"093232c1-8151-4eac-97ec-76c29f228e05","Type":"ContainerDied","Data":"712e3a0139796a4a1434856e06363d8133baf9310624eb20d5916f30f3d4fd86"} Oct 02 09:48:39 crc kubenswrapper[5035]: I1002 09:48:39.246726 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0400439-a652-4472-9f8a-51323ea10780","Type":"ContainerStarted","Data":"ef2b5da4342b0bf896cc7684dd006f53f6cf7648eab850f5ae3f09509a8bb075"} Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.261139 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0400439-a652-4472-9f8a-51323ea10780","Type":"ContainerStarted","Data":"d9bdf1a1a3ca49d754e6c9eb99aea0d38533dac951f71cfc149d4b5840d15815"} Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.662202 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.811302 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-combined-ca-bundle\") pod \"093232c1-8151-4eac-97ec-76c29f228e05\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.811406 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-scripts\") pod \"093232c1-8151-4eac-97ec-76c29f228e05\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.811490 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5k5m\" (UniqueName: \"kubernetes.io/projected/093232c1-8151-4eac-97ec-76c29f228e05-kube-api-access-w5k5m\") pod \"093232c1-8151-4eac-97ec-76c29f228e05\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.811751 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-config-data\") pod \"093232c1-8151-4eac-97ec-76c29f228e05\" (UID: \"093232c1-8151-4eac-97ec-76c29f228e05\") " Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.817773 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-scripts" (OuterVolumeSpecName: "scripts") pod "093232c1-8151-4eac-97ec-76c29f228e05" (UID: "093232c1-8151-4eac-97ec-76c29f228e05"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.817950 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093232c1-8151-4eac-97ec-76c29f228e05-kube-api-access-w5k5m" (OuterVolumeSpecName: "kube-api-access-w5k5m") pod "093232c1-8151-4eac-97ec-76c29f228e05" (UID: "093232c1-8151-4eac-97ec-76c29f228e05"). InnerVolumeSpecName "kube-api-access-w5k5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.844574 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "093232c1-8151-4eac-97ec-76c29f228e05" (UID: "093232c1-8151-4eac-97ec-76c29f228e05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.844822 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-config-data" (OuterVolumeSpecName: "config-data") pod "093232c1-8151-4eac-97ec-76c29f228e05" (UID: "093232c1-8151-4eac-97ec-76c29f228e05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.914481 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5k5m\" (UniqueName: \"kubernetes.io/projected/093232c1-8151-4eac-97ec-76c29f228e05-kube-api-access-w5k5m\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.914549 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.914563 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:40 crc kubenswrapper[5035]: I1002 09:48:40.914573 5035 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/093232c1-8151-4eac-97ec-76c29f228e05-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.287481 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0400439-a652-4472-9f8a-51323ea10780","Type":"ContainerStarted","Data":"f239161fbac3cc497fd63bc81d5526c1482b20aa98b663d3c34d336b21291a8c"} Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.290879 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dsl64" event={"ID":"093232c1-8151-4eac-97ec-76c29f228e05","Type":"ContainerDied","Data":"31c312094181789ed3f53ec924cf351ebb8d3f75444de441784d012a70f1ecd4"} Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.291040 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31c312094181789ed3f53ec924cf351ebb8d3f75444de441784d012a70f1ecd4" Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.290953 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dsl64" Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.518575 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.519794 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4d1e4714-45d9-4930-a186-43210468b573" containerName="nova-scheduler-scheduler" containerID="cri-o://ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5" gracePeriod=30 Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.528737 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.529081 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-log" containerID="cri-o://21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb" gracePeriod=30 Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.529630 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-api" containerID="cri-o://62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803" gracePeriod=30 Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.545713 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.545958 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-log" containerID="cri-o://3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d" gracePeriod=30 Oct 02 09:48:41 crc kubenswrapper[5035]: I1002 09:48:41.546452 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-metadata" containerID="cri-o://0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659" gracePeriod=30 Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.105067 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.240888 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-public-tls-certs\") pod \"c1ebd571-0a0d-473e-b502-b3727de8b617\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.240991 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-config-data\") pod \"c1ebd571-0a0d-473e-b502-b3727de8b617\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.241009 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-combined-ca-bundle\") pod \"c1ebd571-0a0d-473e-b502-b3727de8b617\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.241145 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgl74\" (UniqueName: \"kubernetes.io/projected/c1ebd571-0a0d-473e-b502-b3727de8b617-kube-api-access-rgl74\") pod \"c1ebd571-0a0d-473e-b502-b3727de8b617\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.241191 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1ebd571-0a0d-473e-b502-b3727de8b617-logs\") pod \"c1ebd571-0a0d-473e-b502-b3727de8b617\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.241329 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-internal-tls-certs\") pod \"c1ebd571-0a0d-473e-b502-b3727de8b617\" (UID: \"c1ebd571-0a0d-473e-b502-b3727de8b617\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.241601 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1ebd571-0a0d-473e-b502-b3727de8b617-logs" (OuterVolumeSpecName: "logs") pod "c1ebd571-0a0d-473e-b502-b3727de8b617" (UID: "c1ebd571-0a0d-473e-b502-b3727de8b617"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.241866 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c1ebd571-0a0d-473e-b502-b3727de8b617-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.249714 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1ebd571-0a0d-473e-b502-b3727de8b617-kube-api-access-rgl74" (OuterVolumeSpecName: "kube-api-access-rgl74") pod "c1ebd571-0a0d-473e-b502-b3727de8b617" (UID: "c1ebd571-0a0d-473e-b502-b3727de8b617"). InnerVolumeSpecName "kube-api-access-rgl74". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.283762 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1ebd571-0a0d-473e-b502-b3727de8b617" (UID: "c1ebd571-0a0d-473e-b502-b3727de8b617"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.291678 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-config-data" (OuterVolumeSpecName: "config-data") pod "c1ebd571-0a0d-473e-b502-b3727de8b617" (UID: "c1ebd571-0a0d-473e-b502-b3727de8b617"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.299732 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c1ebd571-0a0d-473e-b502-b3727de8b617" (UID: "c1ebd571-0a0d-473e-b502-b3727de8b617"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.302761 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c1ebd571-0a0d-473e-b502-b3727de8b617" (UID: "c1ebd571-0a0d-473e-b502-b3727de8b617"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.313782 5035 generic.go:334] "Generic (PLEG): container finished" podID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerID="3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d" exitCode=143 Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.313860 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e57a6e9b-12ef-4611-861f-f57215d0edc6","Type":"ContainerDied","Data":"3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d"} Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.319292 5035 generic.go:334] "Generic (PLEG): container finished" podID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerID="62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803" exitCode=0 Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.319339 5035 generic.go:334] "Generic (PLEG): container finished" podID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerID="21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb" exitCode=143 Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.319407 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1ebd571-0a0d-473e-b502-b3727de8b617","Type":"ContainerDied","Data":"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803"} Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.319409 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.319435 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1ebd571-0a0d-473e-b502-b3727de8b617","Type":"ContainerDied","Data":"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb"} Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.319447 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c1ebd571-0a0d-473e-b502-b3727de8b617","Type":"ContainerDied","Data":"2ca4bd645d25ee623ebcf93f265799ea78a41068dbeac583cf9ba0e7d92cfdae"} Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.319463 5035 scope.go:117] "RemoveContainer" containerID="62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.323399 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a0400439-a652-4472-9f8a-51323ea10780","Type":"ContainerStarted","Data":"510e53cb106dbd7e4d5ddd7f6522272427c7fca85a184eac062e2b021b87257e"} Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.324262 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.344894 5035 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.344930 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.344945 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.344959 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgl74\" (UniqueName: \"kubernetes.io/projected/c1ebd571-0a0d-473e-b502-b3727de8b617-kube-api-access-rgl74\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.344970 5035 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1ebd571-0a0d-473e-b502-b3727de8b617-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.349181 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.039055281 podStartE2EDuration="5.349161072s" podCreationTimestamp="2025-10-02 09:48:37 +0000 UTC" firstStartedPulling="2025-10-02 09:48:38.09832491 +0000 UTC m=+1283.454668935" lastFinishedPulling="2025-10-02 09:48:41.408430701 +0000 UTC m=+1286.764774726" observedRunningTime="2025-10-02 09:48:42.342742485 +0000 UTC m=+1287.699086530" watchObservedRunningTime="2025-10-02 09:48:42.349161072 +0000 UTC m=+1287.705505097" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.361143 5035 scope.go:117] "RemoveContainer" containerID="21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.388641 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.403708 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.407442 5035 scope.go:117] "RemoveContainer" containerID="62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803" Oct 02 09:48:42 crc kubenswrapper[5035]: E1002 09:48:42.408649 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803\": container with ID starting with 62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803 not found: ID does not exist" containerID="62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.408686 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803"} err="failed to get container status \"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803\": rpc error: code = NotFound desc = could not find container \"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803\": container with ID starting with 62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803 not found: ID does not exist" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.408710 5035 scope.go:117] "RemoveContainer" containerID="21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb" Oct 02 09:48:42 crc kubenswrapper[5035]: E1002 09:48:42.409474 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb\": container with ID starting with 21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb not found: ID does not exist" containerID="21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.409521 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb"} err="failed to get container status \"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb\": rpc error: code = NotFound desc = could not find container \"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb\": container with ID starting with 21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb not found: ID does not exist" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.409552 5035 scope.go:117] "RemoveContainer" containerID="62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.415296 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803"} err="failed to get container status \"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803\": rpc error: code = NotFound desc = could not find container \"62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803\": container with ID starting with 62fcfafec749d4023a5bdb9af2dada045b100d4e4b91746599e1e8677c32e803 not found: ID does not exist" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.415324 5035 scope.go:117] "RemoveContainer" containerID="21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.416361 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:42 crc kubenswrapper[5035]: E1002 09:48:42.416923 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-log" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.416937 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-log" Oct 02 09:48:42 crc kubenswrapper[5035]: E1002 09:48:42.416977 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-api" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.416982 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-api" Oct 02 09:48:42 crc kubenswrapper[5035]: E1002 09:48:42.417003 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093232c1-8151-4eac-97ec-76c29f228e05" containerName="nova-manage" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.417009 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="093232c1-8151-4eac-97ec-76c29f228e05" containerName="nova-manage" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.417302 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb"} err="failed to get container status \"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb\": rpc error: code = NotFound desc = could not find container \"21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb\": container with ID starting with 21dcdf29ed47547870b8e1c029a3088f987e21d081413dc2ee6ec4b412de18eb not found: ID does not exist" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.417693 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-api" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.417721 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" containerName="nova-api-log" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.417730 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="093232c1-8151-4eac-97ec-76c29f228e05" containerName="nova-manage" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.418782 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.427250 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.427820 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.428488 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.429815 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.550600 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kllxg\" (UniqueName: \"kubernetes.io/projected/75df5c05-833c-4534-8a47-2130520e96ac-kube-api-access-kllxg\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.550679 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75df5c05-833c-4534-8a47-2130520e96ac-logs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.550714 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.550758 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.550812 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-config-data\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.550839 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.652597 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75df5c05-833c-4534-8a47-2130520e96ac-logs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.652974 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.653031 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.653097 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-config-data\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.653130 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.653183 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kllxg\" (UniqueName: \"kubernetes.io/projected/75df5c05-833c-4534-8a47-2130520e96ac-kube-api-access-kllxg\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.654928 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75df5c05-833c-4534-8a47-2130520e96ac-logs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.660112 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-config-data\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.674114 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-public-tls-certs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.674250 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.674549 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75df5c05-833c-4534-8a47-2130520e96ac-internal-tls-certs\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.678220 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kllxg\" (UniqueName: \"kubernetes.io/projected/75df5c05-833c-4534-8a47-2130520e96ac-kube-api-access-kllxg\") pod \"nova-api-0\" (UID: \"75df5c05-833c-4534-8a47-2130520e96ac\") " pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.749093 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.766296 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.865814 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mnnr\" (UniqueName: \"kubernetes.io/projected/4d1e4714-45d9-4930-a186-43210468b573-kube-api-access-7mnnr\") pod \"4d1e4714-45d9-4930-a186-43210468b573\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.865904 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-config-data\") pod \"4d1e4714-45d9-4930-a186-43210468b573\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.865969 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-combined-ca-bundle\") pod \"4d1e4714-45d9-4930-a186-43210468b573\" (UID: \"4d1e4714-45d9-4930-a186-43210468b573\") " Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.873218 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d1e4714-45d9-4930-a186-43210468b573-kube-api-access-7mnnr" (OuterVolumeSpecName: "kube-api-access-7mnnr") pod "4d1e4714-45d9-4930-a186-43210468b573" (UID: "4d1e4714-45d9-4930-a186-43210468b573"). InnerVolumeSpecName "kube-api-access-7mnnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.894404 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-config-data" (OuterVolumeSpecName: "config-data") pod "4d1e4714-45d9-4930-a186-43210468b573" (UID: "4d1e4714-45d9-4930-a186-43210468b573"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.915592 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d1e4714-45d9-4930-a186-43210468b573" (UID: "4d1e4714-45d9-4930-a186-43210468b573"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.968335 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mnnr\" (UniqueName: \"kubernetes.io/projected/4d1e4714-45d9-4930-a186-43210468b573-kube-api-access-7mnnr\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.968815 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:42 crc kubenswrapper[5035]: I1002 09:48:42.968829 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1e4714-45d9-4930-a186-43210468b573-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.249309 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 09:48:43 crc kubenswrapper[5035]: W1002 09:48:43.251128 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75df5c05_833c_4534_8a47_2130520e96ac.slice/crio-c5073f8ed16180b2f459bbc5d0e98257a045d355ad7424cd28078fc13b5916b2 WatchSource:0}: Error finding container c5073f8ed16180b2f459bbc5d0e98257a045d355ad7424cd28078fc13b5916b2: Status 404 returned error can't find the container with id c5073f8ed16180b2f459bbc5d0e98257a045d355ad7424cd28078fc13b5916b2 Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.339200 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75df5c05-833c-4534-8a47-2130520e96ac","Type":"ContainerStarted","Data":"c5073f8ed16180b2f459bbc5d0e98257a045d355ad7424cd28078fc13b5916b2"} Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.341401 5035 generic.go:334] "Generic (PLEG): container finished" podID="4d1e4714-45d9-4930-a186-43210468b573" containerID="ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5" exitCode=0 Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.341449 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d1e4714-45d9-4930-a186-43210468b573","Type":"ContainerDied","Data":"ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5"} Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.341468 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d1e4714-45d9-4930-a186-43210468b573","Type":"ContainerDied","Data":"22485c98714e4081d54ed8c66ec8a44be9958a7ed39b3f8cc2085f7242ba62b9"} Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.341484 5035 scope.go:117] "RemoveContainer" containerID="ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.341479 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.383958 5035 scope.go:117] "RemoveContainer" containerID="ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5" Oct 02 09:48:43 crc kubenswrapper[5035]: E1002 09:48:43.384515 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5\": container with ID starting with ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5 not found: ID does not exist" containerID="ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.384605 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5"} err="failed to get container status \"ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5\": rpc error: code = NotFound desc = could not find container \"ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5\": container with ID starting with ae846ed82f55a202e9a85892fe8060bff232d2f19b0fdfb2ab8574f08e4a9cf5 not found: ID does not exist" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.404689 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.419631 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.449196 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:43 crc kubenswrapper[5035]: E1002 09:48:43.450327 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d1e4714-45d9-4930-a186-43210468b573" containerName="nova-scheduler-scheduler" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.450375 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d1e4714-45d9-4930-a186-43210468b573" containerName="nova-scheduler-scheduler" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.450770 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d1e4714-45d9-4930-a186-43210468b573" containerName="nova-scheduler-scheduler" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.451950 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.456003 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.457937 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.585266 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c61dad9-d722-4d17-958f-594d3d87ca14-config-data\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.585366 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c61dad9-d722-4d17-958f-594d3d87ca14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.585666 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9z6b\" (UniqueName: \"kubernetes.io/projected/7c61dad9-d722-4d17-958f-594d3d87ca14-kube-api-access-b9z6b\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.687676 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9z6b\" (UniqueName: \"kubernetes.io/projected/7c61dad9-d722-4d17-958f-594d3d87ca14-kube-api-access-b9z6b\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.687786 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c61dad9-d722-4d17-958f-594d3d87ca14-config-data\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.687813 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c61dad9-d722-4d17-958f-594d3d87ca14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.692219 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7c61dad9-d722-4d17-958f-594d3d87ca14-config-data\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.692983 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c61dad9-d722-4d17-958f-594d3d87ca14-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.706289 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9z6b\" (UniqueName: \"kubernetes.io/projected/7c61dad9-d722-4d17-958f-594d3d87ca14-kube-api-access-b9z6b\") pod \"nova-scheduler-0\" (UID: \"7c61dad9-d722-4d17-958f-594d3d87ca14\") " pod="openstack/nova-scheduler-0" Oct 02 09:48:43 crc kubenswrapper[5035]: I1002 09:48:43.783963 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.174609 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d1e4714-45d9-4930-a186-43210468b573" path="/var/lib/kubelet/pods/4d1e4714-45d9-4930-a186-43210468b573/volumes" Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.175205 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1ebd571-0a0d-473e-b502-b3727de8b617" path="/var/lib/kubelet/pods/c1ebd571-0a0d-473e-b502-b3727de8b617/volumes" Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.238378 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 09:48:44 crc kubenswrapper[5035]: W1002 09:48:44.244240 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c61dad9_d722_4d17_958f_594d3d87ca14.slice/crio-60f8a1f0ec7f4e3124e6368d72d5373530fb852fce802162ad3f9d2e7ee22896 WatchSource:0}: Error finding container 60f8a1f0ec7f4e3124e6368d72d5373530fb852fce802162ad3f9d2e7ee22896: Status 404 returned error can't find the container with id 60f8a1f0ec7f4e3124e6368d72d5373530fb852fce802162ad3f9d2e7ee22896 Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.353598 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7c61dad9-d722-4d17-958f-594d3d87ca14","Type":"ContainerStarted","Data":"60f8a1f0ec7f4e3124e6368d72d5373530fb852fce802162ad3f9d2e7ee22896"} Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.355708 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75df5c05-833c-4534-8a47-2130520e96ac","Type":"ContainerStarted","Data":"71e80a0807863113272f7ef5d2985a66617587641eca142d56fbf9fba1d3f89d"} Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.355753 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"75df5c05-833c-4534-8a47-2130520e96ac","Type":"ContainerStarted","Data":"0967f8979e6bbb0129d12b23bc9367f0a81760935993bd14af6f6fa419c94d76"} Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.375945 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.375927579 podStartE2EDuration="2.375927579s" podCreationTimestamp="2025-10-02 09:48:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:44.374291402 +0000 UTC m=+1289.730635437" watchObservedRunningTime="2025-10-02 09:48:44.375927579 +0000 UTC m=+1289.732271594" Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.673672 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:54656->10.217.0.204:8775: read: connection reset by peer" Oct 02 09:48:44 crc kubenswrapper[5035]: I1002 09:48:44.673727 5035 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": read tcp 10.217.0.2:54660->10.217.0.204:8775: read: connection reset by peer" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.206203 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.320554 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7rh7\" (UniqueName: \"kubernetes.io/projected/e57a6e9b-12ef-4611-861f-f57215d0edc6-kube-api-access-g7rh7\") pod \"e57a6e9b-12ef-4611-861f-f57215d0edc6\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.320741 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-combined-ca-bundle\") pod \"e57a6e9b-12ef-4611-861f-f57215d0edc6\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.320820 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-config-data\") pod \"e57a6e9b-12ef-4611-861f-f57215d0edc6\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.320870 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57a6e9b-12ef-4611-861f-f57215d0edc6-logs\") pod \"e57a6e9b-12ef-4611-861f-f57215d0edc6\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.320898 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-nova-metadata-tls-certs\") pod \"e57a6e9b-12ef-4611-861f-f57215d0edc6\" (UID: \"e57a6e9b-12ef-4611-861f-f57215d0edc6\") " Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.322568 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e57a6e9b-12ef-4611-861f-f57215d0edc6-logs" (OuterVolumeSpecName: "logs") pod "e57a6e9b-12ef-4611-861f-f57215d0edc6" (UID: "e57a6e9b-12ef-4611-861f-f57215d0edc6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.328475 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e57a6e9b-12ef-4611-861f-f57215d0edc6-kube-api-access-g7rh7" (OuterVolumeSpecName: "kube-api-access-g7rh7") pod "e57a6e9b-12ef-4611-861f-f57215d0edc6" (UID: "e57a6e9b-12ef-4611-861f-f57215d0edc6"). InnerVolumeSpecName "kube-api-access-g7rh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.357296 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e57a6e9b-12ef-4611-861f-f57215d0edc6" (UID: "e57a6e9b-12ef-4611-861f-f57215d0edc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.362641 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-config-data" (OuterVolumeSpecName: "config-data") pod "e57a6e9b-12ef-4611-861f-f57215d0edc6" (UID: "e57a6e9b-12ef-4611-861f-f57215d0edc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.371373 5035 generic.go:334] "Generic (PLEG): container finished" podID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerID="0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659" exitCode=0 Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.371469 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.371547 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e57a6e9b-12ef-4611-861f-f57215d0edc6","Type":"ContainerDied","Data":"0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659"} Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.371642 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e57a6e9b-12ef-4611-861f-f57215d0edc6","Type":"ContainerDied","Data":"dc44c589879a4b332744650db2625730f5912abc0444ab66d1453ab531615eca"} Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.371672 5035 scope.go:117] "RemoveContainer" containerID="0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.373499 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7c61dad9-d722-4d17-958f-594d3d87ca14","Type":"ContainerStarted","Data":"a684fb7d61f9d31b666e898a25fd3aec1f4787f227f5b2909237a53364a24ed2"} Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.407754 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.407729825 podStartE2EDuration="2.407729825s" podCreationTimestamp="2025-10-02 09:48:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:45.39657544 +0000 UTC m=+1290.752919465" watchObservedRunningTime="2025-10-02 09:48:45.407729825 +0000 UTC m=+1290.764073850" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.412915 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e57a6e9b-12ef-4611-861f-f57215d0edc6" (UID: "e57a6e9b-12ef-4611-861f-f57215d0edc6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.423065 5035 scope.go:117] "RemoveContainer" containerID="3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.424023 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.424052 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.424063 5035 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e57a6e9b-12ef-4611-861f-f57215d0edc6-logs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.424075 5035 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e57a6e9b-12ef-4611-861f-f57215d0edc6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.424088 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7rh7\" (UniqueName: \"kubernetes.io/projected/e57a6e9b-12ef-4611-861f-f57215d0edc6-kube-api-access-g7rh7\") on node \"crc\" DevicePath \"\"" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.450219 5035 scope.go:117] "RemoveContainer" containerID="0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659" Oct 02 09:48:45 crc kubenswrapper[5035]: E1002 09:48:45.451457 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659\": container with ID starting with 0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659 not found: ID does not exist" containerID="0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.451502 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659"} err="failed to get container status \"0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659\": rpc error: code = NotFound desc = could not find container \"0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659\": container with ID starting with 0e4ee8f3b7911561094e33de24e2716be05ae8943917ed7ae7579c168a420659 not found: ID does not exist" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.451542 5035 scope.go:117] "RemoveContainer" containerID="3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d" Oct 02 09:48:45 crc kubenswrapper[5035]: E1002 09:48:45.451920 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d\": container with ID starting with 3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d not found: ID does not exist" containerID="3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.451940 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d"} err="failed to get container status \"3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d\": rpc error: code = NotFound desc = could not find container \"3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d\": container with ID starting with 3303d4bf100caa2ef8e87fd1d1d6da112591bbbb599ec0449eadebd5990e773d not found: ID does not exist" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.706615 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.729121 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.754559 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:48:45 crc kubenswrapper[5035]: E1002 09:48:45.754998 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-metadata" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.755016 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-metadata" Oct 02 09:48:45 crc kubenswrapper[5035]: E1002 09:48:45.755055 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-log" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.755063 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-log" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.755265 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-log" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.755280 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" containerName="nova-metadata-metadata" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.756244 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.760127 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.760180 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.774203 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.934236 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.934362 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-config-data\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.934422 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm6lk\" (UniqueName: \"kubernetes.io/projected/b803caac-28bd-4f36-b701-e6774f8b83bd-kube-api-access-xm6lk\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.934507 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:45 crc kubenswrapper[5035]: I1002 09:48:45.934551 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b803caac-28bd-4f36-b701-e6774f8b83bd-logs\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.035775 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.035881 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-config-data\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.035918 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm6lk\" (UniqueName: \"kubernetes.io/projected/b803caac-28bd-4f36-b701-e6774f8b83bd-kube-api-access-xm6lk\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.035979 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.035997 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b803caac-28bd-4f36-b701-e6774f8b83bd-logs\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.036374 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b803caac-28bd-4f36-b701-e6774f8b83bd-logs\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.042589 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.043847 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.055625 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b803caac-28bd-4f36-b701-e6774f8b83bd-config-data\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.068307 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm6lk\" (UniqueName: \"kubernetes.io/projected/b803caac-28bd-4f36-b701-e6774f8b83bd-kube-api-access-xm6lk\") pod \"nova-metadata-0\" (UID: \"b803caac-28bd-4f36-b701-e6774f8b83bd\") " pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.077606 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.198836 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e57a6e9b-12ef-4611-861f-f57215d0edc6" path="/var/lib/kubelet/pods/e57a6e9b-12ef-4611-861f-f57215d0edc6/volumes" Oct 02 09:48:46 crc kubenswrapper[5035]: I1002 09:48:46.373675 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 09:48:47 crc kubenswrapper[5035]: I1002 09:48:47.400378 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b803caac-28bd-4f36-b701-e6774f8b83bd","Type":"ContainerStarted","Data":"5e72f29ff715a6052d09525a4cc44796b0f1c08f88ad19a3006a6b552e1881c7"} Oct 02 09:48:47 crc kubenswrapper[5035]: I1002 09:48:47.400965 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b803caac-28bd-4f36-b701-e6774f8b83bd","Type":"ContainerStarted","Data":"35e9a363e43a13bc7e7469fc44d9cd6a2f127b87dae72013623763c19a69286b"} Oct 02 09:48:47 crc kubenswrapper[5035]: I1002 09:48:47.400979 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b803caac-28bd-4f36-b701-e6774f8b83bd","Type":"ContainerStarted","Data":"0105f6bf290573a529681186bf9529e51663ddb9d21eaac6bb379b4851d70210"} Oct 02 09:48:47 crc kubenswrapper[5035]: I1002 09:48:47.458702 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.458683826 podStartE2EDuration="2.458683826s" podCreationTimestamp="2025-10-02 09:48:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:48:47.458471959 +0000 UTC m=+1292.814815984" watchObservedRunningTime="2025-10-02 09:48:47.458683826 +0000 UTC m=+1292.815027851" Oct 02 09:48:48 crc kubenswrapper[5035]: I1002 09:48:48.784643 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 09:48:51 crc kubenswrapper[5035]: I1002 09:48:51.059169 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Oct 02 09:48:51 crc kubenswrapper[5035]: I1002 09:48:51.078318 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 09:48:51 crc kubenswrapper[5035]: I1002 09:48:51.078418 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 09:48:52 crc kubenswrapper[5035]: I1002 09:48:52.751036 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 09:48:52 crc kubenswrapper[5035]: I1002 09:48:52.751369 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 09:48:53 crc kubenswrapper[5035]: I1002 09:48:53.762768 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75df5c05-833c-4534-8a47-2130520e96ac" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 09:48:53 crc kubenswrapper[5035]: I1002 09:48:53.762767 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="75df5c05-833c-4534-8a47-2130520e96ac" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 09:48:53 crc kubenswrapper[5035]: I1002 09:48:53.784371 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 09:48:53 crc kubenswrapper[5035]: I1002 09:48:53.812333 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 09:48:54 crc kubenswrapper[5035]: I1002 09:48:54.492666 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 09:48:56 crc kubenswrapper[5035]: I1002 09:48:56.078747 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 09:48:56 crc kubenswrapper[5035]: I1002 09:48:56.078805 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 09:48:57 crc kubenswrapper[5035]: I1002 09:48:57.090813 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b803caac-28bd-4f36-b701-e6774f8b83bd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 09:48:57 crc kubenswrapper[5035]: I1002 09:48:57.090835 5035 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b803caac-28bd-4f36-b701-e6774f8b83bd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 09:49:02 crc kubenswrapper[5035]: I1002 09:49:02.758643 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 09:49:02 crc kubenswrapper[5035]: I1002 09:49:02.759516 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 09:49:02 crc kubenswrapper[5035]: I1002 09:49:02.766921 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 09:49:02 crc kubenswrapper[5035]: I1002 09:49:02.767876 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 09:49:03 crc kubenswrapper[5035]: I1002 09:49:03.563083 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 09:49:03 crc kubenswrapper[5035]: I1002 09:49:03.571692 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 09:49:06 crc kubenswrapper[5035]: I1002 09:49:06.084158 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 09:49:06 crc kubenswrapper[5035]: I1002 09:49:06.086030 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 09:49:06 crc kubenswrapper[5035]: I1002 09:49:06.089676 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 09:49:06 crc kubenswrapper[5035]: I1002 09:49:06.596644 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 09:49:07 crc kubenswrapper[5035]: I1002 09:49:07.636385 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 09:49:25 crc kubenswrapper[5035]: I1002 09:49:25.538357 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:49:25 crc kubenswrapper[5035]: I1002 09:49:25.539090 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:49:55 crc kubenswrapper[5035]: I1002 09:49:55.538636 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:49:55 crc kubenswrapper[5035]: I1002 09:49:55.539149 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:50:25 crc kubenswrapper[5035]: I1002 09:50:25.537679 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:50:25 crc kubenswrapper[5035]: I1002 09:50:25.538726 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:50:25 crc kubenswrapper[5035]: I1002 09:50:25.539169 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:50:25 crc kubenswrapper[5035]: I1002 09:50:25.539987 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6c5f2097a0a83c78bab17bcdbccb738feed82e531c37ccc7b2955b0ca92c739"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:50:25 crc kubenswrapper[5035]: I1002 09:50:25.540058 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://e6c5f2097a0a83c78bab17bcdbccb738feed82e531c37ccc7b2955b0ca92c739" gracePeriod=600 Oct 02 09:50:26 crc kubenswrapper[5035]: I1002 09:50:26.406603 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="e6c5f2097a0a83c78bab17bcdbccb738feed82e531c37ccc7b2955b0ca92c739" exitCode=0 Oct 02 09:50:26 crc kubenswrapper[5035]: I1002 09:50:26.406707 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"e6c5f2097a0a83c78bab17bcdbccb738feed82e531c37ccc7b2955b0ca92c739"} Oct 02 09:50:26 crc kubenswrapper[5035]: I1002 09:50:26.407157 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363"} Oct 02 09:50:26 crc kubenswrapper[5035]: I1002 09:50:26.407179 5035 scope.go:117] "RemoveContainer" containerID="db9ffa0950502b83608fcd73436cfa37aeaa8a721d02dc03e8e35488a5442a33" Oct 02 09:50:34 crc kubenswrapper[5035]: I1002 09:50:34.803365 5035 scope.go:117] "RemoveContainer" containerID="03f45f06468cc4881431ccb308d8a34e6a4960701f3592cf3dcf729cf1921bdf" Oct 02 09:50:34 crc kubenswrapper[5035]: I1002 09:50:34.841364 5035 scope.go:117] "RemoveContainer" containerID="48f226ce96c69da3e41b507332f66fe23ed15441d0c4ac7aa86053a91d748990" Oct 02 09:50:34 crc kubenswrapper[5035]: I1002 09:50:34.890791 5035 scope.go:117] "RemoveContainer" containerID="4ab336006445281cc0f01128884b576e880092ba1dc49158b120b76b24913468" Oct 02 09:50:34 crc kubenswrapper[5035]: I1002 09:50:34.927199 5035 scope.go:117] "RemoveContainer" containerID="98958a9bc11d70d8847c9327d8c86fa88c1fcc75899eae7980e734aca897beb8" Oct 02 09:51:35 crc kubenswrapper[5035]: I1002 09:51:35.031988 5035 scope.go:117] "RemoveContainer" containerID="3f51f48e5f9e423c6395355aebd9732b5fcc9b082f94afe2a777f3f5890a3cea" Oct 02 09:51:35 crc kubenswrapper[5035]: I1002 09:51:35.079185 5035 scope.go:117] "RemoveContainer" containerID="804befe089b471e947159d43e917b186670d15949200f28417981b3bfffae808" Oct 02 09:51:35 crc kubenswrapper[5035]: I1002 09:51:35.113716 5035 scope.go:117] "RemoveContainer" containerID="f2c89a8d7b3ee8367edfb6af778e14bc3505cf972e0f8764ea0d989c7b3e5d34" Oct 02 09:51:35 crc kubenswrapper[5035]: I1002 09:51:35.137334 5035 scope.go:117] "RemoveContainer" containerID="173049a936e4990ec816ab9bbdb1fbeca88ec082a9aa7593cdea4d1076f4c304" Oct 02 09:51:35 crc kubenswrapper[5035]: I1002 09:51:35.161210 5035 scope.go:117] "RemoveContainer" containerID="c602845ee5438d0817f63a5900578511888dcf10b25a37336986d1f8931e2f29" Oct 02 09:51:35 crc kubenswrapper[5035]: I1002 09:51:35.197678 5035 scope.go:117] "RemoveContainer" containerID="6ea426aa25725e5a82f3a3f4772279450441598ebdc4296ba755846b29b4fd6a" Oct 02 09:51:35 crc kubenswrapper[5035]: I1002 09:51:35.229545 5035 scope.go:117] "RemoveContainer" containerID="95640f4b076c18d15a63d7cfac95605fe0fce07eb6470486af1f04e53256acac" Oct 02 09:51:36 crc kubenswrapper[5035]: I1002 09:51:36.955505 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9n6nq"] Oct 02 09:51:36 crc kubenswrapper[5035]: I1002 09:51:36.959341 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:36 crc kubenswrapper[5035]: I1002 09:51:36.970400 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9n6nq"] Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.062161 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8gls\" (UniqueName: \"kubernetes.io/projected/c56707e6-3ba4-4fd4-acb5-069682f41c32-kube-api-access-m8gls\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.062640 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-catalog-content\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.062791 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-utilities\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.164382 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-catalog-content\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.164455 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-utilities\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.164515 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8gls\" (UniqueName: \"kubernetes.io/projected/c56707e6-3ba4-4fd4-acb5-069682f41c32-kube-api-access-m8gls\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.165036 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-catalog-content\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.165300 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-utilities\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.186517 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8gls\" (UniqueName: \"kubernetes.io/projected/c56707e6-3ba4-4fd4-acb5-069682f41c32-kube-api-access-m8gls\") pod \"redhat-operators-9n6nq\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.284962 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:37 crc kubenswrapper[5035]: I1002 09:51:37.775875 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9n6nq"] Oct 02 09:51:37 crc kubenswrapper[5035]: W1002 09:51:37.787106 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc56707e6_3ba4_4fd4_acb5_069682f41c32.slice/crio-93d535f1a6eb1ad3e1a730d160ef93050191511d3e02eb844fcbd1658bebd301 WatchSource:0}: Error finding container 93d535f1a6eb1ad3e1a730d160ef93050191511d3e02eb844fcbd1658bebd301: Status 404 returned error can't find the container with id 93d535f1a6eb1ad3e1a730d160ef93050191511d3e02eb844fcbd1658bebd301 Oct 02 09:51:38 crc kubenswrapper[5035]: I1002 09:51:38.168397 5035 generic.go:334] "Generic (PLEG): container finished" podID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerID="53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8" exitCode=0 Oct 02 09:51:38 crc kubenswrapper[5035]: I1002 09:51:38.177304 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n6nq" event={"ID":"c56707e6-3ba4-4fd4-acb5-069682f41c32","Type":"ContainerDied","Data":"53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8"} Oct 02 09:51:38 crc kubenswrapper[5035]: I1002 09:51:38.177365 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n6nq" event={"ID":"c56707e6-3ba4-4fd4-acb5-069682f41c32","Type":"ContainerStarted","Data":"93d535f1a6eb1ad3e1a730d160ef93050191511d3e02eb844fcbd1658bebd301"} Oct 02 09:51:40 crc kubenswrapper[5035]: I1002 09:51:40.203940 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n6nq" event={"ID":"c56707e6-3ba4-4fd4-acb5-069682f41c32","Type":"ContainerStarted","Data":"9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc"} Oct 02 09:51:41 crc kubenswrapper[5035]: I1002 09:51:41.216588 5035 generic.go:334] "Generic (PLEG): container finished" podID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerID="9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc" exitCode=0 Oct 02 09:51:41 crc kubenswrapper[5035]: I1002 09:51:41.216635 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n6nq" event={"ID":"c56707e6-3ba4-4fd4-acb5-069682f41c32","Type":"ContainerDied","Data":"9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc"} Oct 02 09:51:42 crc kubenswrapper[5035]: I1002 09:51:42.234918 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n6nq" event={"ID":"c56707e6-3ba4-4fd4-acb5-069682f41c32","Type":"ContainerStarted","Data":"a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97"} Oct 02 09:51:42 crc kubenswrapper[5035]: I1002 09:51:42.259333 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9n6nq" podStartSLOduration=2.519066252 podStartE2EDuration="6.259308599s" podCreationTimestamp="2025-10-02 09:51:36 +0000 UTC" firstStartedPulling="2025-10-02 09:51:38.170635072 +0000 UTC m=+1463.526979097" lastFinishedPulling="2025-10-02 09:51:41.910877419 +0000 UTC m=+1467.267221444" observedRunningTime="2025-10-02 09:51:42.252996204 +0000 UTC m=+1467.609340229" watchObservedRunningTime="2025-10-02 09:51:42.259308599 +0000 UTC m=+1467.615652624" Oct 02 09:51:47 crc kubenswrapper[5035]: I1002 09:51:47.285196 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:47 crc kubenswrapper[5035]: I1002 09:51:47.285850 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:47 crc kubenswrapper[5035]: I1002 09:51:47.337172 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:48 crc kubenswrapper[5035]: I1002 09:51:48.338967 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:48 crc kubenswrapper[5035]: I1002 09:51:48.411932 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9n6nq"] Oct 02 09:51:50 crc kubenswrapper[5035]: I1002 09:51:50.324865 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9n6nq" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="registry-server" containerID="cri-o://a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97" gracePeriod=2 Oct 02 09:51:50 crc kubenswrapper[5035]: I1002 09:51:50.813743 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:50 crc kubenswrapper[5035]: I1002 09:51:50.963998 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8gls\" (UniqueName: \"kubernetes.io/projected/c56707e6-3ba4-4fd4-acb5-069682f41c32-kube-api-access-m8gls\") pod \"c56707e6-3ba4-4fd4-acb5-069682f41c32\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " Oct 02 09:51:50 crc kubenswrapper[5035]: I1002 09:51:50.964739 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-utilities\") pod \"c56707e6-3ba4-4fd4-acb5-069682f41c32\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " Oct 02 09:51:50 crc kubenswrapper[5035]: I1002 09:51:50.965267 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-catalog-content\") pod \"c56707e6-3ba4-4fd4-acb5-069682f41c32\" (UID: \"c56707e6-3ba4-4fd4-acb5-069682f41c32\") " Oct 02 09:51:50 crc kubenswrapper[5035]: I1002 09:51:50.965853 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-utilities" (OuterVolumeSpecName: "utilities") pod "c56707e6-3ba4-4fd4-acb5-069682f41c32" (UID: "c56707e6-3ba4-4fd4-acb5-069682f41c32"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:51:50 crc kubenswrapper[5035]: I1002 09:51:50.973059 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c56707e6-3ba4-4fd4-acb5-069682f41c32-kube-api-access-m8gls" (OuterVolumeSpecName: "kube-api-access-m8gls") pod "c56707e6-3ba4-4fd4-acb5-069682f41c32" (UID: "c56707e6-3ba4-4fd4-acb5-069682f41c32"). InnerVolumeSpecName "kube-api-access-m8gls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.049546 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c56707e6-3ba4-4fd4-acb5-069682f41c32" (UID: "c56707e6-3ba4-4fd4-acb5-069682f41c32"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.067509 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.067587 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8gls\" (UniqueName: \"kubernetes.io/projected/c56707e6-3ba4-4fd4-acb5-069682f41c32-kube-api-access-m8gls\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.067609 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c56707e6-3ba4-4fd4-acb5-069682f41c32-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.339239 5035 generic.go:334] "Generic (PLEG): container finished" podID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerID="a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97" exitCode=0 Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.339294 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9n6nq" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.339294 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n6nq" event={"ID":"c56707e6-3ba4-4fd4-acb5-069682f41c32","Type":"ContainerDied","Data":"a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97"} Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.339419 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9n6nq" event={"ID":"c56707e6-3ba4-4fd4-acb5-069682f41c32","Type":"ContainerDied","Data":"93d535f1a6eb1ad3e1a730d160ef93050191511d3e02eb844fcbd1658bebd301"} Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.339443 5035 scope.go:117] "RemoveContainer" containerID="a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.364768 5035 scope.go:117] "RemoveContainer" containerID="9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.380720 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9n6nq"] Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.385494 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9n6nq"] Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.409841 5035 scope.go:117] "RemoveContainer" containerID="53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.451028 5035 scope.go:117] "RemoveContainer" containerID="a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97" Oct 02 09:51:51 crc kubenswrapper[5035]: E1002 09:51:51.451540 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97\": container with ID starting with a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97 not found: ID does not exist" containerID="a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.451789 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97"} err="failed to get container status \"a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97\": rpc error: code = NotFound desc = could not find container \"a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97\": container with ID starting with a9a763e38ffa84b21b9d55c898618752f9265f1609f6833026b41dfd7366de97 not found: ID does not exist" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.451960 5035 scope.go:117] "RemoveContainer" containerID="9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc" Oct 02 09:51:51 crc kubenswrapper[5035]: E1002 09:51:51.452408 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc\": container with ID starting with 9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc not found: ID does not exist" containerID="9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.452494 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc"} err="failed to get container status \"9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc\": rpc error: code = NotFound desc = could not find container \"9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc\": container with ID starting with 9c1a08f891e56515550d2152b1326dd2cfc47de1cee6d79808d2872f097c66fc not found: ID does not exist" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.452573 5035 scope.go:117] "RemoveContainer" containerID="53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8" Oct 02 09:51:51 crc kubenswrapper[5035]: E1002 09:51:51.452857 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8\": container with ID starting with 53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8 not found: ID does not exist" containerID="53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8" Oct 02 09:51:51 crc kubenswrapper[5035]: I1002 09:51:51.452883 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8"} err="failed to get container status \"53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8\": rpc error: code = NotFound desc = could not find container \"53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8\": container with ID starting with 53b5d2a98269349e16334ab453b6dfcba30c58e8c16a2972d7e4e9e9f8913df8 not found: ID does not exist" Oct 02 09:51:52 crc kubenswrapper[5035]: I1002 09:51:52.175104 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" path="/var/lib/kubelet/pods/c56707e6-3ba4-4fd4-acb5-069682f41c32/volumes" Oct 02 09:52:25 crc kubenswrapper[5035]: I1002 09:52:25.537768 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:52:25 crc kubenswrapper[5035]: I1002 09:52:25.538338 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.200365 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zxw6c"] Oct 02 09:52:27 crc kubenswrapper[5035]: E1002 09:52:27.201166 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="extract-utilities" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.201184 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="extract-utilities" Oct 02 09:52:27 crc kubenswrapper[5035]: E1002 09:52:27.201223 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="extract-content" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.201233 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="extract-content" Oct 02 09:52:27 crc kubenswrapper[5035]: E1002 09:52:27.201258 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="registry-server" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.201267 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="registry-server" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.201469 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="c56707e6-3ba4-4fd4-acb5-069682f41c32" containerName="registry-server" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.203150 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.215407 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxw6c"] Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.290768 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-utilities\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.290916 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2jmp\" (UniqueName: \"kubernetes.io/projected/95110aad-918d-4419-a1c2-67b33193cc3b-kube-api-access-v2jmp\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.290980 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-catalog-content\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.392625 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2jmp\" (UniqueName: \"kubernetes.io/projected/95110aad-918d-4419-a1c2-67b33193cc3b-kube-api-access-v2jmp\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.392761 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-catalog-content\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.392805 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-utilities\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.393344 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-utilities\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.393424 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-catalog-content\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.422287 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2jmp\" (UniqueName: \"kubernetes.io/projected/95110aad-918d-4419-a1c2-67b33193cc3b-kube-api-access-v2jmp\") pod \"redhat-marketplace-zxw6c\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:27 crc kubenswrapper[5035]: I1002 09:52:27.529649 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:28 crc kubenswrapper[5035]: I1002 09:52:28.004635 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxw6c"] Oct 02 09:52:28 crc kubenswrapper[5035]: I1002 09:52:28.716947 5035 generic.go:334] "Generic (PLEG): container finished" podID="95110aad-918d-4419-a1c2-67b33193cc3b" containerID="cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a" exitCode=0 Oct 02 09:52:28 crc kubenswrapper[5035]: I1002 09:52:28.717036 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxw6c" event={"ID":"95110aad-918d-4419-a1c2-67b33193cc3b","Type":"ContainerDied","Data":"cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a"} Oct 02 09:52:28 crc kubenswrapper[5035]: I1002 09:52:28.717259 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxw6c" event={"ID":"95110aad-918d-4419-a1c2-67b33193cc3b","Type":"ContainerStarted","Data":"02374547c6103640f990369ae6f18dac6c233cf4bc0ae7de0ca2ea09a1f35a64"} Oct 02 09:52:29 crc kubenswrapper[5035]: I1002 09:52:29.731041 5035 generic.go:334] "Generic (PLEG): container finished" podID="95110aad-918d-4419-a1c2-67b33193cc3b" containerID="95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8" exitCode=0 Oct 02 09:52:29 crc kubenswrapper[5035]: I1002 09:52:29.731173 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxw6c" event={"ID":"95110aad-918d-4419-a1c2-67b33193cc3b","Type":"ContainerDied","Data":"95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8"} Oct 02 09:52:30 crc kubenswrapper[5035]: I1002 09:52:30.746850 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxw6c" event={"ID":"95110aad-918d-4419-a1c2-67b33193cc3b","Type":"ContainerStarted","Data":"b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299"} Oct 02 09:52:30 crc kubenswrapper[5035]: I1002 09:52:30.770226 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zxw6c" podStartSLOduration=2.1878089689999998 podStartE2EDuration="3.770206249s" podCreationTimestamp="2025-10-02 09:52:27 +0000 UTC" firstStartedPulling="2025-10-02 09:52:28.720092857 +0000 UTC m=+1514.076436882" lastFinishedPulling="2025-10-02 09:52:30.302490137 +0000 UTC m=+1515.658834162" observedRunningTime="2025-10-02 09:52:30.768312193 +0000 UTC m=+1516.124656218" watchObservedRunningTime="2025-10-02 09:52:30.770206249 +0000 UTC m=+1516.126550264" Oct 02 09:52:35 crc kubenswrapper[5035]: I1002 09:52:35.345968 5035 scope.go:117] "RemoveContainer" containerID="50a5c12ef817b3257d4f0e3edf6c3ec5dad480722699ffb96a4fda1c6d53c2a1" Oct 02 09:52:35 crc kubenswrapper[5035]: I1002 09:52:35.380255 5035 scope.go:117] "RemoveContainer" containerID="a10d5e846c8ea3f7112b3a156244cc8d64d57eeb3b7d582c28802531b8587bbe" Oct 02 09:52:37 crc kubenswrapper[5035]: I1002 09:52:37.529995 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:37 crc kubenswrapper[5035]: I1002 09:52:37.530310 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:37 crc kubenswrapper[5035]: I1002 09:52:37.578483 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:37 crc kubenswrapper[5035]: I1002 09:52:37.860761 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:37 crc kubenswrapper[5035]: I1002 09:52:37.913324 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxw6c"] Oct 02 09:52:39 crc kubenswrapper[5035]: I1002 09:52:39.834773 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zxw6c" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="registry-server" containerID="cri-o://b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299" gracePeriod=2 Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.289683 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.354831 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2jmp\" (UniqueName: \"kubernetes.io/projected/95110aad-918d-4419-a1c2-67b33193cc3b-kube-api-access-v2jmp\") pod \"95110aad-918d-4419-a1c2-67b33193cc3b\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.355035 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-catalog-content\") pod \"95110aad-918d-4419-a1c2-67b33193cc3b\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.355120 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-utilities\") pod \"95110aad-918d-4419-a1c2-67b33193cc3b\" (UID: \"95110aad-918d-4419-a1c2-67b33193cc3b\") " Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.356266 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-utilities" (OuterVolumeSpecName: "utilities") pod "95110aad-918d-4419-a1c2-67b33193cc3b" (UID: "95110aad-918d-4419-a1c2-67b33193cc3b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.363361 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95110aad-918d-4419-a1c2-67b33193cc3b-kube-api-access-v2jmp" (OuterVolumeSpecName: "kube-api-access-v2jmp") pod "95110aad-918d-4419-a1c2-67b33193cc3b" (UID: "95110aad-918d-4419-a1c2-67b33193cc3b"). InnerVolumeSpecName "kube-api-access-v2jmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.372214 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95110aad-918d-4419-a1c2-67b33193cc3b" (UID: "95110aad-918d-4419-a1c2-67b33193cc3b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.457721 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.457766 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2jmp\" (UniqueName: \"kubernetes.io/projected/95110aad-918d-4419-a1c2-67b33193cc3b-kube-api-access-v2jmp\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.457779 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95110aad-918d-4419-a1c2-67b33193cc3b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.847866 5035 generic.go:334] "Generic (PLEG): container finished" podID="95110aad-918d-4419-a1c2-67b33193cc3b" containerID="b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299" exitCode=0 Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.847921 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxw6c" event={"ID":"95110aad-918d-4419-a1c2-67b33193cc3b","Type":"ContainerDied","Data":"b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299"} Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.848749 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zxw6c" event={"ID":"95110aad-918d-4419-a1c2-67b33193cc3b","Type":"ContainerDied","Data":"02374547c6103640f990369ae6f18dac6c233cf4bc0ae7de0ca2ea09a1f35a64"} Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.848785 5035 scope.go:117] "RemoveContainer" containerID="b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.848212 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zxw6c" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.873345 5035 scope.go:117] "RemoveContainer" containerID="95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.904694 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxw6c"] Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.909745 5035 scope.go:117] "RemoveContainer" containerID="cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.917212 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zxw6c"] Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.954569 5035 scope.go:117] "RemoveContainer" containerID="b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299" Oct 02 09:52:40 crc kubenswrapper[5035]: E1002 09:52:40.955041 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299\": container with ID starting with b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299 not found: ID does not exist" containerID="b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.955074 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299"} err="failed to get container status \"b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299\": rpc error: code = NotFound desc = could not find container \"b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299\": container with ID starting with b3ec145abe5fc6be152ca7f1a63d8ba2998ea39d5b6feee8bf280e4334dbc299 not found: ID does not exist" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.955097 5035 scope.go:117] "RemoveContainer" containerID="95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8" Oct 02 09:52:40 crc kubenswrapper[5035]: E1002 09:52:40.955853 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8\": container with ID starting with 95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8 not found: ID does not exist" containerID="95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.955921 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8"} err="failed to get container status \"95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8\": rpc error: code = NotFound desc = could not find container \"95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8\": container with ID starting with 95e0a14a72999be7696232bbce4a89feba4f10a99793fd2d3ac0fff362202fc8 not found: ID does not exist" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.955967 5035 scope.go:117] "RemoveContainer" containerID="cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a" Oct 02 09:52:40 crc kubenswrapper[5035]: E1002 09:52:40.956442 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a\": container with ID starting with cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a not found: ID does not exist" containerID="cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a" Oct 02 09:52:40 crc kubenswrapper[5035]: I1002 09:52:40.956520 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a"} err="failed to get container status \"cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a\": rpc error: code = NotFound desc = could not find container \"cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a\": container with ID starting with cd7ee223b462805f6595d57953d6f0c1c025a1f4a9a9ad6feadae9c8ce3be10a not found: ID does not exist" Oct 02 09:52:42 crc kubenswrapper[5035]: I1002 09:52:42.176444 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" path="/var/lib/kubelet/pods/95110aad-918d-4419-a1c2-67b33193cc3b/volumes" Oct 02 09:52:55 crc kubenswrapper[5035]: I1002 09:52:55.537472 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:52:55 crc kubenswrapper[5035]: I1002 09:52:55.538050 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.421520 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hm726"] Oct 02 09:52:59 crc kubenswrapper[5035]: E1002 09:52:59.422193 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="extract-content" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.422205 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="extract-content" Oct 02 09:52:59 crc kubenswrapper[5035]: E1002 09:52:59.422222 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="extract-utilities" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.422228 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="extract-utilities" Oct 02 09:52:59 crc kubenswrapper[5035]: E1002 09:52:59.422255 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="registry-server" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.422263 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="registry-server" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.422441 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="95110aad-918d-4419-a1c2-67b33193cc3b" containerName="registry-server" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.424817 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.445810 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hm726"] Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.450060 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bzzw\" (UniqueName: \"kubernetes.io/projected/5936ee5c-aa54-45ac-961d-d0032457d7d5-kube-api-access-8bzzw\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.450143 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-catalog-content\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.450179 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-utilities\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.552393 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bzzw\" (UniqueName: \"kubernetes.io/projected/5936ee5c-aa54-45ac-961d-d0032457d7d5-kube-api-access-8bzzw\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.552506 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-catalog-content\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.552636 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-utilities\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.553048 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-catalog-content\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.553232 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-utilities\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.574991 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bzzw\" (UniqueName: \"kubernetes.io/projected/5936ee5c-aa54-45ac-961d-d0032457d7d5-kube-api-access-8bzzw\") pod \"community-operators-hm726\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " pod="openshift-marketplace/community-operators-hm726" Oct 02 09:52:59 crc kubenswrapper[5035]: I1002 09:52:59.752587 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:53:00 crc kubenswrapper[5035]: I1002 09:53:00.271018 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hm726"] Oct 02 09:53:01 crc kubenswrapper[5035]: I1002 09:53:01.058980 5035 generic.go:334] "Generic (PLEG): container finished" podID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerID="6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430" exitCode=0 Oct 02 09:53:01 crc kubenswrapper[5035]: I1002 09:53:01.059144 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hm726" event={"ID":"5936ee5c-aa54-45ac-961d-d0032457d7d5","Type":"ContainerDied","Data":"6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430"} Oct 02 09:53:01 crc kubenswrapper[5035]: I1002 09:53:01.059320 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hm726" event={"ID":"5936ee5c-aa54-45ac-961d-d0032457d7d5","Type":"ContainerStarted","Data":"1a4128bc037342c845fd071329603afbbfac0af086842a8ef66c82c0e0a677ef"} Oct 02 09:53:04 crc kubenswrapper[5035]: I1002 09:53:04.106302 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hm726" event={"ID":"5936ee5c-aa54-45ac-961d-d0032457d7d5","Type":"ContainerStarted","Data":"b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56"} Oct 02 09:53:05 crc kubenswrapper[5035]: I1002 09:53:05.118118 5035 generic.go:334] "Generic (PLEG): container finished" podID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerID="b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56" exitCode=0 Oct 02 09:53:05 crc kubenswrapper[5035]: I1002 09:53:05.118186 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hm726" event={"ID":"5936ee5c-aa54-45ac-961d-d0032457d7d5","Type":"ContainerDied","Data":"b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56"} Oct 02 09:53:06 crc kubenswrapper[5035]: I1002 09:53:06.130135 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hm726" event={"ID":"5936ee5c-aa54-45ac-961d-d0032457d7d5","Type":"ContainerStarted","Data":"b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78"} Oct 02 09:53:06 crc kubenswrapper[5035]: I1002 09:53:06.157840 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hm726" podStartSLOduration=2.66682461 podStartE2EDuration="7.157812683s" podCreationTimestamp="2025-10-02 09:52:59 +0000 UTC" firstStartedPulling="2025-10-02 09:53:01.061058928 +0000 UTC m=+1546.417402953" lastFinishedPulling="2025-10-02 09:53:05.552047011 +0000 UTC m=+1550.908391026" observedRunningTime="2025-10-02 09:53:06.14847271 +0000 UTC m=+1551.504816735" watchObservedRunningTime="2025-10-02 09:53:06.157812683 +0000 UTC m=+1551.514156708" Oct 02 09:53:09 crc kubenswrapper[5035]: I1002 09:53:09.753324 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:53:09 crc kubenswrapper[5035]: I1002 09:53:09.754182 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:53:09 crc kubenswrapper[5035]: I1002 09:53:09.802410 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:53:10 crc kubenswrapper[5035]: I1002 09:53:10.242195 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:53:10 crc kubenswrapper[5035]: I1002 09:53:10.301355 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hm726"] Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.188431 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hm726" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="registry-server" containerID="cri-o://b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78" gracePeriod=2 Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.655351 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.730344 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bzzw\" (UniqueName: \"kubernetes.io/projected/5936ee5c-aa54-45ac-961d-d0032457d7d5-kube-api-access-8bzzw\") pod \"5936ee5c-aa54-45ac-961d-d0032457d7d5\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.730675 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-utilities\") pod \"5936ee5c-aa54-45ac-961d-d0032457d7d5\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.730708 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-catalog-content\") pod \"5936ee5c-aa54-45ac-961d-d0032457d7d5\" (UID: \"5936ee5c-aa54-45ac-961d-d0032457d7d5\") " Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.731438 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-utilities" (OuterVolumeSpecName: "utilities") pod "5936ee5c-aa54-45ac-961d-d0032457d7d5" (UID: "5936ee5c-aa54-45ac-961d-d0032457d7d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.736991 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5936ee5c-aa54-45ac-961d-d0032457d7d5-kube-api-access-8bzzw" (OuterVolumeSpecName: "kube-api-access-8bzzw") pod "5936ee5c-aa54-45ac-961d-d0032457d7d5" (UID: "5936ee5c-aa54-45ac-961d-d0032457d7d5"). InnerVolumeSpecName "kube-api-access-8bzzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.791603 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5936ee5c-aa54-45ac-961d-d0032457d7d5" (UID: "5936ee5c-aa54-45ac-961d-d0032457d7d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.832840 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bzzw\" (UniqueName: \"kubernetes.io/projected/5936ee5c-aa54-45ac-961d-d0032457d7d5-kube-api-access-8bzzw\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.832884 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:12 crc kubenswrapper[5035]: I1002 09:53:12.832900 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5936ee5c-aa54-45ac-961d-d0032457d7d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.204049 5035 generic.go:334] "Generic (PLEG): container finished" podID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerID="b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78" exitCode=0 Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.204072 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hm726" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.204095 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hm726" event={"ID":"5936ee5c-aa54-45ac-961d-d0032457d7d5","Type":"ContainerDied","Data":"b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78"} Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.204894 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hm726" event={"ID":"5936ee5c-aa54-45ac-961d-d0032457d7d5","Type":"ContainerDied","Data":"1a4128bc037342c845fd071329603afbbfac0af086842a8ef66c82c0e0a677ef"} Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.205054 5035 scope.go:117] "RemoveContainer" containerID="b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.235832 5035 scope.go:117] "RemoveContainer" containerID="b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.256715 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hm726"] Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.266377 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hm726"] Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.288936 5035 scope.go:117] "RemoveContainer" containerID="6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.339164 5035 scope.go:117] "RemoveContainer" containerID="b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78" Oct 02 09:53:13 crc kubenswrapper[5035]: E1002 09:53:13.339735 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78\": container with ID starting with b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78 not found: ID does not exist" containerID="b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.339773 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78"} err="failed to get container status \"b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78\": rpc error: code = NotFound desc = could not find container \"b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78\": container with ID starting with b9e922c72860be11195ebee6844171f2932e25619d96f5a516a3663539332a78 not found: ID does not exist" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.339799 5035 scope.go:117] "RemoveContainer" containerID="b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56" Oct 02 09:53:13 crc kubenswrapper[5035]: E1002 09:53:13.340085 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56\": container with ID starting with b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56 not found: ID does not exist" containerID="b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.340116 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56"} err="failed to get container status \"b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56\": rpc error: code = NotFound desc = could not find container \"b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56\": container with ID starting with b2712beeb1a05c196820f7f133cd25e2c0f9b3cb9948c068d8d80a6426ec0f56 not found: ID does not exist" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.340134 5035 scope.go:117] "RemoveContainer" containerID="6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430" Oct 02 09:53:13 crc kubenswrapper[5035]: E1002 09:53:13.340606 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430\": container with ID starting with 6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430 not found: ID does not exist" containerID="6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430" Oct 02 09:53:13 crc kubenswrapper[5035]: I1002 09:53:13.340639 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430"} err="failed to get container status \"6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430\": rpc error: code = NotFound desc = could not find container \"6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430\": container with ID starting with 6723666e71533377e040222cf832cf91628b19d19c5b57020ca15e588a765430 not found: ID does not exist" Oct 02 09:53:14 crc kubenswrapper[5035]: I1002 09:53:14.178248 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" path="/var/lib/kubelet/pods/5936ee5c-aa54-45ac-961d-d0032457d7d5/volumes" Oct 02 09:53:25 crc kubenswrapper[5035]: I1002 09:53:25.538833 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:53:25 crc kubenswrapper[5035]: I1002 09:53:25.539395 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:53:25 crc kubenswrapper[5035]: I1002 09:53:25.539463 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 09:53:25 crc kubenswrapper[5035]: I1002 09:53:25.540604 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:53:25 crc kubenswrapper[5035]: I1002 09:53:25.540700 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" gracePeriod=600 Oct 02 09:53:25 crc kubenswrapper[5035]: E1002 09:53:25.665914 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:53:26 crc kubenswrapper[5035]: I1002 09:53:26.374759 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" exitCode=0 Oct 02 09:53:26 crc kubenswrapper[5035]: I1002 09:53:26.374819 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363"} Oct 02 09:53:26 crc kubenswrapper[5035]: I1002 09:53:26.375118 5035 scope.go:117] "RemoveContainer" containerID="e6c5f2097a0a83c78bab17bcdbccb738feed82e531c37ccc7b2955b0ca92c739" Oct 02 09:53:26 crc kubenswrapper[5035]: I1002 09:53:26.375877 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:53:26 crc kubenswrapper[5035]: E1002 09:53:26.376234 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:53:35 crc kubenswrapper[5035]: I1002 09:53:35.476579 5035 scope.go:117] "RemoveContainer" containerID="19fae54dd1bd848b4159124317b70000e34344a77f8bb966d012b4464d769df5" Oct 02 09:53:35 crc kubenswrapper[5035]: I1002 09:53:35.499428 5035 scope.go:117] "RemoveContainer" containerID="3e2a2d08fddfab03d09c2ac4491ed93b33809cbfe5f53c15ed51531ecf4c3cfa" Oct 02 09:53:35 crc kubenswrapper[5035]: I1002 09:53:35.528352 5035 scope.go:117] "RemoveContainer" containerID="622091f5b215247208af56a8c6910d3577bc3c313113973b8976dbbd8fb501a4" Oct 02 09:53:35 crc kubenswrapper[5035]: I1002 09:53:35.558560 5035 scope.go:117] "RemoveContainer" containerID="ee9375d91aab70bed1379563b9054b67667a95a24627db26e986c25bac1cad5e" Oct 02 09:53:35 crc kubenswrapper[5035]: I1002 09:53:35.603470 5035 scope.go:117] "RemoveContainer" containerID="c3fea9636911d0bdcc0510bd44047092d0c3d3bb1f8594c1003bdafbcd3fb3f2" Oct 02 09:53:35 crc kubenswrapper[5035]: I1002 09:53:35.646331 5035 scope.go:117] "RemoveContainer" containerID="6af8c80800b309d2da419734305b2d58faf447ba1d49c041eb7614b6652f54e1" Oct 02 09:53:40 crc kubenswrapper[5035]: I1002 09:53:40.163284 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:53:40 crc kubenswrapper[5035]: E1002 09:53:40.164142 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:53:55 crc kubenswrapper[5035]: I1002 09:53:55.164414 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:53:55 crc kubenswrapper[5035]: E1002 09:53:55.165109 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:54:01 crc kubenswrapper[5035]: I1002 09:54:01.046629 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bqd4v"] Oct 02 09:54:01 crc kubenswrapper[5035]: I1002 09:54:01.057809 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wc7vd"] Oct 02 09:54:01 crc kubenswrapper[5035]: I1002 09:54:01.065159 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-d5gmk"] Oct 02 09:54:01 crc kubenswrapper[5035]: I1002 09:54:01.073191 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bqd4v"] Oct 02 09:54:01 crc kubenswrapper[5035]: I1002 09:54:01.081053 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-d5gmk"] Oct 02 09:54:01 crc kubenswrapper[5035]: I1002 09:54:01.090012 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wc7vd"] Oct 02 09:54:02 crc kubenswrapper[5035]: I1002 09:54:02.176381 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb09fa9f-7359-46a9-baf8-758474e12535" path="/var/lib/kubelet/pods/cb09fa9f-7359-46a9-baf8-758474e12535/volumes" Oct 02 09:54:02 crc kubenswrapper[5035]: I1002 09:54:02.177626 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e56670c7-7dcd-4294-b852-cf1397b892f7" path="/var/lib/kubelet/pods/e56670c7-7dcd-4294-b852-cf1397b892f7/volumes" Oct 02 09:54:02 crc kubenswrapper[5035]: I1002 09:54:02.178619 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed3ec4f1-c604-42e1-bffd-255bf7586994" path="/var/lib/kubelet/pods/ed3ec4f1-c604-42e1-bffd-255bf7586994/volumes" Oct 02 09:54:08 crc kubenswrapper[5035]: I1002 09:54:08.163103 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:54:08 crc kubenswrapper[5035]: E1002 09:54:08.164002 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:54:13 crc kubenswrapper[5035]: I1002 09:54:13.029601 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2a86-account-create-rvm4s"] Oct 02 09:54:13 crc kubenswrapper[5035]: I1002 09:54:13.040247 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-11a0-account-create-xf5dg"] Oct 02 09:54:13 crc kubenswrapper[5035]: I1002 09:54:13.050896 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1d44-account-create-tgq25"] Oct 02 09:54:13 crc kubenswrapper[5035]: I1002 09:54:13.059243 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-11a0-account-create-xf5dg"] Oct 02 09:54:13 crc kubenswrapper[5035]: I1002 09:54:13.068781 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2a86-account-create-rvm4s"] Oct 02 09:54:13 crc kubenswrapper[5035]: I1002 09:54:13.076402 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1d44-account-create-tgq25"] Oct 02 09:54:14 crc kubenswrapper[5035]: I1002 09:54:14.174004 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="533bb1e5-5f6a-4f87-a671-1af95bb013c4" path="/var/lib/kubelet/pods/533bb1e5-5f6a-4f87-a671-1af95bb013c4/volumes" Oct 02 09:54:14 crc kubenswrapper[5035]: I1002 09:54:14.174606 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="772fa808-a60d-4cbf-a6a8-b7e7e2345ad1" path="/var/lib/kubelet/pods/772fa808-a60d-4cbf-a6a8-b7e7e2345ad1/volumes" Oct 02 09:54:14 crc kubenswrapper[5035]: I1002 09:54:14.175077 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fbf41c1-35f8-4689-92ce-5338a740e524" path="/var/lib/kubelet/pods/8fbf41c1-35f8-4689-92ce-5338a740e524/volumes" Oct 02 09:54:18 crc kubenswrapper[5035]: I1002 09:54:18.027811 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-jmk82"] Oct 02 09:54:18 crc kubenswrapper[5035]: I1002 09:54:18.036345 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-jmk82"] Oct 02 09:54:18 crc kubenswrapper[5035]: I1002 09:54:18.209081 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76e148c5-87f2-440a-adae-0d3cc1f67f2a" path="/var/lib/kubelet/pods/76e148c5-87f2-440a-adae-0d3cc1f67f2a/volumes" Oct 02 09:54:19 crc kubenswrapper[5035]: I1002 09:54:19.025803 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-tkjv5"] Oct 02 09:54:19 crc kubenswrapper[5035]: I1002 09:54:19.034313 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-gjbv9"] Oct 02 09:54:19 crc kubenswrapper[5035]: I1002 09:54:19.043624 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-gjbv9"] Oct 02 09:54:19 crc kubenswrapper[5035]: I1002 09:54:19.051179 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-tkjv5"] Oct 02 09:54:20 crc kubenswrapper[5035]: I1002 09:54:20.176524 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68687a35-da55-497b-a7fa-0fe15b039d15" path="/var/lib/kubelet/pods/68687a35-da55-497b-a7fa-0fe15b039d15/volumes" Oct 02 09:54:20 crc kubenswrapper[5035]: I1002 09:54:20.177243 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e29f134c-e329-4296-b491-101f71a255a0" path="/var/lib/kubelet/pods/e29f134c-e329-4296-b491-101f71a255a0/volumes" Oct 02 09:54:22 crc kubenswrapper[5035]: I1002 09:54:22.163484 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:54:22 crc kubenswrapper[5035]: E1002 09:54:22.164491 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:54:33 crc kubenswrapper[5035]: I1002 09:54:33.162776 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:54:33 crc kubenswrapper[5035]: E1002 09:54:33.163739 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:54:34 crc kubenswrapper[5035]: I1002 09:54:34.045738 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8a80-account-create-666mg"] Oct 02 09:54:34 crc kubenswrapper[5035]: I1002 09:54:34.053879 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6096-account-create-vxgpc"] Oct 02 09:54:34 crc kubenswrapper[5035]: I1002 09:54:34.061353 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6096-account-create-vxgpc"] Oct 02 09:54:34 crc kubenswrapper[5035]: I1002 09:54:34.069727 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8a80-account-create-666mg"] Oct 02 09:54:34 crc kubenswrapper[5035]: I1002 09:54:34.176669 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37719aef-3ac7-486b-a59b-4acd555afcc1" path="/var/lib/kubelet/pods/37719aef-3ac7-486b-a59b-4acd555afcc1/volumes" Oct 02 09:54:34 crc kubenswrapper[5035]: I1002 09:54:34.177204 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50486824-cd3b-489e-986f-d5107331b8b1" path="/var/lib/kubelet/pods/50486824-cd3b-489e-986f-d5107331b8b1/volumes" Oct 02 09:54:35 crc kubenswrapper[5035]: I1002 09:54:35.756179 5035 scope.go:117] "RemoveContainer" containerID="2b3daca74aad2f3efdfff4772d258e1591628504c5da8ea5ab47a8bf55fa16aa" Oct 02 09:54:35 crc kubenswrapper[5035]: I1002 09:54:35.782399 5035 scope.go:117] "RemoveContainer" containerID="c09ce0b256e5d90e50314bc158d09f0eddede663e178107504e385cc41220f52" Oct 02 09:54:35 crc kubenswrapper[5035]: I1002 09:54:35.800342 5035 scope.go:117] "RemoveContainer" containerID="4653b9a0a9c73a2a75d3f3ade0365b3d1a7f78880a01063517758e0f30840ebc" Oct 02 09:54:35 crc kubenswrapper[5035]: I1002 09:54:35.822877 5035 scope.go:117] "RemoveContainer" containerID="7ffeb550cf987e795f5d2e0401dbb2dd58531bdaf6b5a851e9b4618f6ee17357" Oct 02 09:54:35 crc kubenswrapper[5035]: I1002 09:54:35.876179 5035 scope.go:117] "RemoveContainer" containerID="d26dbb8808e679c5fa7c9faac4b005d667e058c6c1f127995e662021738fa8bf" Oct 02 09:54:35 crc kubenswrapper[5035]: I1002 09:54:35.918916 5035 scope.go:117] "RemoveContainer" containerID="c971a4caed2be13cdbd97fc3bcd7270df0f68140db6230049b17ea142569a286" Oct 02 09:54:35 crc kubenswrapper[5035]: I1002 09:54:35.959360 5035 scope.go:117] "RemoveContainer" containerID="4ccde4beae1e00ad4b185cfebf60366ec2f45f85fa1fb6d555f0521ea26b7c18" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.070759 5035 scope.go:117] "RemoveContainer" containerID="0d1f9166aab340a416355f48608eb5a3413a9049b66064a4ba94e116a83aa557" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.097879 5035 scope.go:117] "RemoveContainer" containerID="5b5131abcbb5bc8ef131d86e4a1556e66b7d39654c483641b0dfce473cb447dc" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.119388 5035 scope.go:117] "RemoveContainer" containerID="27a080b5795472921ae4e21f43bb5869372bb174c20776b899ade1b0191e9cc2" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.138507 5035 scope.go:117] "RemoveContainer" containerID="47ce2159e127d801ff299e7bf0640995b0ebb35ccd9d56e89cf3b1c4476ba02c" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.164910 5035 scope.go:117] "RemoveContainer" containerID="3eb3fea218d76ef05e18df0c8a55add715a9fd458d3194af7aaad8e5c2bc5925" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.188220 5035 scope.go:117] "RemoveContainer" containerID="e547550edd7974f43c6afa0784465d4329e93f5ef3b963dcfd3381d05c8cb767" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.210117 5035 scope.go:117] "RemoveContainer" containerID="35746bdb68bb8babda9b723a6b9201cd3834f755fb27faa1404263889d05b2e0" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.230166 5035 scope.go:117] "RemoveContainer" containerID="20dc0a2ea25fa97bb7ae731fe42920cb0582ac2484f3e902bee72485975fcc90" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.260773 5035 scope.go:117] "RemoveContainer" containerID="5383078a22cfae40ed3d838e37ee6a049538d61cfb7b6391bfb86beaa7240088" Oct 02 09:54:36 crc kubenswrapper[5035]: I1002 09:54:36.283855 5035 scope.go:117] "RemoveContainer" containerID="49c6a0043f09fa0e68c4968b1ce55de721d9a1a073a4f8e976c48b283d70d338" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.030135 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-7c81-account-create-nk5lf"] Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.038669 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-7c81-account-create-nk5lf"] Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.183309 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0155abf0-b42c-4a52-adf4-cef5dfc9c1b6" path="/var/lib/kubelet/pods/0155abf0-b42c-4a52-adf4-cef5dfc9c1b6/volumes" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.493788 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6r2xl"] Oct 02 09:54:38 crc kubenswrapper[5035]: E1002 09:54:38.494167 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="registry-server" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.494184 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="registry-server" Oct 02 09:54:38 crc kubenswrapper[5035]: E1002 09:54:38.494205 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="extract-utilities" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.494212 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="extract-utilities" Oct 02 09:54:38 crc kubenswrapper[5035]: E1002 09:54:38.494225 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="extract-content" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.494234 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="extract-content" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.494453 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5936ee5c-aa54-45ac-961d-d0032457d7d5" containerName="registry-server" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.495869 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.514927 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6r2xl"] Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.618052 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz9k2\" (UniqueName: \"kubernetes.io/projected/54958464-af3c-45c5-b4bd-bc1a07f6f758-kube-api-access-fz9k2\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.618673 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-catalog-content\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.618838 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-utilities\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.721864 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz9k2\" (UniqueName: \"kubernetes.io/projected/54958464-af3c-45c5-b4bd-bc1a07f6f758-kube-api-access-fz9k2\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.721945 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-catalog-content\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.722044 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-utilities\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.722630 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-utilities\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.722751 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-catalog-content\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.745624 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz9k2\" (UniqueName: \"kubernetes.io/projected/54958464-af3c-45c5-b4bd-bc1a07f6f758-kube-api-access-fz9k2\") pod \"certified-operators-6r2xl\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:38 crc kubenswrapper[5035]: I1002 09:54:38.815833 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:39 crc kubenswrapper[5035]: I1002 09:54:39.045844 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-zhwgl"] Oct 02 09:54:39 crc kubenswrapper[5035]: I1002 09:54:39.053045 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-zhwgl"] Oct 02 09:54:39 crc kubenswrapper[5035]: I1002 09:54:39.338191 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6r2xl"] Oct 02 09:54:40 crc kubenswrapper[5035]: I1002 09:54:40.162449 5035 generic.go:334] "Generic (PLEG): container finished" podID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerID="c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4" exitCode=0 Oct 02 09:54:40 crc kubenswrapper[5035]: I1002 09:54:40.164590 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:54:40 crc kubenswrapper[5035]: I1002 09:54:40.179219 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bf48931-dee2-4f51-a3dc-d69a4604064b" path="/var/lib/kubelet/pods/5bf48931-dee2-4f51-a3dc-d69a4604064b/volumes" Oct 02 09:54:40 crc kubenswrapper[5035]: I1002 09:54:40.181482 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6r2xl" event={"ID":"54958464-af3c-45c5-b4bd-bc1a07f6f758","Type":"ContainerDied","Data":"c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4"} Oct 02 09:54:40 crc kubenswrapper[5035]: I1002 09:54:40.181521 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6r2xl" event={"ID":"54958464-af3c-45c5-b4bd-bc1a07f6f758","Type":"ContainerStarted","Data":"8db763bd255bcb424fbf219a38817e61e8d1ec4fdda38d117cffb9c5237d805c"} Oct 02 09:54:41 crc kubenswrapper[5035]: I1002 09:54:41.177773 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6r2xl" event={"ID":"54958464-af3c-45c5-b4bd-bc1a07f6f758","Type":"ContainerStarted","Data":"3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd"} Oct 02 09:54:42 crc kubenswrapper[5035]: I1002 09:54:42.188305 5035 generic.go:334] "Generic (PLEG): container finished" podID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerID="3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd" exitCode=0 Oct 02 09:54:42 crc kubenswrapper[5035]: I1002 09:54:42.188364 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6r2xl" event={"ID":"54958464-af3c-45c5-b4bd-bc1a07f6f758","Type":"ContainerDied","Data":"3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd"} Oct 02 09:54:43 crc kubenswrapper[5035]: I1002 09:54:43.204555 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6r2xl" event={"ID":"54958464-af3c-45c5-b4bd-bc1a07f6f758","Type":"ContainerStarted","Data":"48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b"} Oct 02 09:54:43 crc kubenswrapper[5035]: I1002 09:54:43.230277 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6r2xl" podStartSLOduration=2.769279745 podStartE2EDuration="5.230256152s" podCreationTimestamp="2025-10-02 09:54:38 +0000 UTC" firstStartedPulling="2025-10-02 09:54:40.164281983 +0000 UTC m=+1645.520626008" lastFinishedPulling="2025-10-02 09:54:42.62525839 +0000 UTC m=+1647.981602415" observedRunningTime="2025-10-02 09:54:43.223715319 +0000 UTC m=+1648.580059344" watchObservedRunningTime="2025-10-02 09:54:43.230256152 +0000 UTC m=+1648.586600177" Oct 02 09:54:45 crc kubenswrapper[5035]: I1002 09:54:45.031123 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-db-create-mqjfl"] Oct 02 09:54:45 crc kubenswrapper[5035]: I1002 09:54:45.039191 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-db-create-mqjfl"] Oct 02 09:54:46 crc kubenswrapper[5035]: I1002 09:54:46.174041 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa5bbaa2-0524-45c3-9643-6bc946e45870" path="/var/lib/kubelet/pods/aa5bbaa2-0524-45c3-9643-6bc946e45870/volumes" Oct 02 09:54:48 crc kubenswrapper[5035]: I1002 09:54:48.164067 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:54:48 crc kubenswrapper[5035]: E1002 09:54:48.164690 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:54:48 crc kubenswrapper[5035]: I1002 09:54:48.816235 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:48 crc kubenswrapper[5035]: I1002 09:54:48.816863 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:48 crc kubenswrapper[5035]: I1002 09:54:48.869159 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:49 crc kubenswrapper[5035]: I1002 09:54:49.316730 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:49 crc kubenswrapper[5035]: I1002 09:54:49.373797 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6r2xl"] Oct 02 09:54:50 crc kubenswrapper[5035]: I1002 09:54:50.028775 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-hbb7f"] Oct 02 09:54:50 crc kubenswrapper[5035]: I1002 09:54:50.037164 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-hbb7f"] Oct 02 09:54:50 crc kubenswrapper[5035]: I1002 09:54:50.175789 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f3e489-f6ce-4126-adcb-764afd7e7159" path="/var/lib/kubelet/pods/f6f3e489-f6ce-4126-adcb-764afd7e7159/volumes" Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.274997 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6r2xl" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="registry-server" containerID="cri-o://48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b" gracePeriod=2 Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.724566 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.787723 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-catalog-content\") pod \"54958464-af3c-45c5-b4bd-bc1a07f6f758\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.788178 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-utilities\") pod \"54958464-af3c-45c5-b4bd-bc1a07f6f758\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.788303 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz9k2\" (UniqueName: \"kubernetes.io/projected/54958464-af3c-45c5-b4bd-bc1a07f6f758-kube-api-access-fz9k2\") pod \"54958464-af3c-45c5-b4bd-bc1a07f6f758\" (UID: \"54958464-af3c-45c5-b4bd-bc1a07f6f758\") " Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.789962 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-utilities" (OuterVolumeSpecName: "utilities") pod "54958464-af3c-45c5-b4bd-bc1a07f6f758" (UID: "54958464-af3c-45c5-b4bd-bc1a07f6f758"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.794405 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54958464-af3c-45c5-b4bd-bc1a07f6f758-kube-api-access-fz9k2" (OuterVolumeSpecName: "kube-api-access-fz9k2") pod "54958464-af3c-45c5-b4bd-bc1a07f6f758" (UID: "54958464-af3c-45c5-b4bd-bc1a07f6f758"). InnerVolumeSpecName "kube-api-access-fz9k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.835837 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54958464-af3c-45c5-b4bd-bc1a07f6f758" (UID: "54958464-af3c-45c5-b4bd-bc1a07f6f758"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.890856 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz9k2\" (UniqueName: \"kubernetes.io/projected/54958464-af3c-45c5-b4bd-bc1a07f6f758-kube-api-access-fz9k2\") on node \"crc\" DevicePath \"\"" Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.890939 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:54:51 crc kubenswrapper[5035]: I1002 09:54:51.890952 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54958464-af3c-45c5-b4bd-bc1a07f6f758-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.286444 5035 generic.go:334] "Generic (PLEG): container finished" podID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerID="48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b" exitCode=0 Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.286497 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6r2xl" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.286514 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6r2xl" event={"ID":"54958464-af3c-45c5-b4bd-bc1a07f6f758","Type":"ContainerDied","Data":"48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b"} Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.287611 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6r2xl" event={"ID":"54958464-af3c-45c5-b4bd-bc1a07f6f758","Type":"ContainerDied","Data":"8db763bd255bcb424fbf219a38817e61e8d1ec4fdda38d117cffb9c5237d805c"} Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.287636 5035 scope.go:117] "RemoveContainer" containerID="48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.313821 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6r2xl"] Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.320603 5035 scope.go:117] "RemoveContainer" containerID="3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.328476 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6r2xl"] Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.341296 5035 scope.go:117] "RemoveContainer" containerID="c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.389817 5035 scope.go:117] "RemoveContainer" containerID="48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b" Oct 02 09:54:52 crc kubenswrapper[5035]: E1002 09:54:52.390390 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b\": container with ID starting with 48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b not found: ID does not exist" containerID="48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.390438 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b"} err="failed to get container status \"48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b\": rpc error: code = NotFound desc = could not find container \"48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b\": container with ID starting with 48c0e4bd60097fc9e902a968936d93c31f599d6619cd2ccf0c60a7acf665fe6b not found: ID does not exist" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.390482 5035 scope.go:117] "RemoveContainer" containerID="3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd" Oct 02 09:54:52 crc kubenswrapper[5035]: E1002 09:54:52.390907 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd\": container with ID starting with 3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd not found: ID does not exist" containerID="3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.390947 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd"} err="failed to get container status \"3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd\": rpc error: code = NotFound desc = could not find container \"3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd\": container with ID starting with 3622857c15010429f044f3805d5da8a97ef66b84368fee8433a24d298f4c0dbd not found: ID does not exist" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.390974 5035 scope.go:117] "RemoveContainer" containerID="c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4" Oct 02 09:54:52 crc kubenswrapper[5035]: E1002 09:54:52.391398 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4\": container with ID starting with c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4 not found: ID does not exist" containerID="c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4" Oct 02 09:54:52 crc kubenswrapper[5035]: I1002 09:54:52.391426 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4"} err="failed to get container status \"c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4\": rpc error: code = NotFound desc = could not find container \"c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4\": container with ID starting with c18f097de2a9e689b78d0c0d65c367810f05a84431ceb8bd901a2396275f4ee4 not found: ID does not exist" Oct 02 09:54:54 crc kubenswrapper[5035]: I1002 09:54:54.175584 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" path="/var/lib/kubelet/pods/54958464-af3c-45c5-b4bd-bc1a07f6f758/volumes" Oct 02 09:55:01 crc kubenswrapper[5035]: I1002 09:55:01.163694 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:55:01 crc kubenswrapper[5035]: E1002 09:55:01.164484 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:55:12 crc kubenswrapper[5035]: I1002 09:55:12.034065 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-8c07-account-create-9t9l8"] Oct 02 09:55:12 crc kubenswrapper[5035]: I1002 09:55:12.042384 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-8c07-account-create-9t9l8"] Oct 02 09:55:12 crc kubenswrapper[5035]: I1002 09:55:12.163650 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:55:12 crc kubenswrapper[5035]: E1002 09:55:12.163899 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:55:12 crc kubenswrapper[5035]: I1002 09:55:12.188157 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2887e3dc-2301-42c6-ac2d-725adefd5fcf" path="/var/lib/kubelet/pods/2887e3dc-2301-42c6-ac2d-725adefd5fcf/volumes" Oct 02 09:55:19 crc kubenswrapper[5035]: I1002 09:55:19.033858 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xb6lf"] Oct 02 09:55:19 crc kubenswrapper[5035]: I1002 09:55:19.043712 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-knpjm"] Oct 02 09:55:19 crc kubenswrapper[5035]: I1002 09:55:19.054287 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xb6lf"] Oct 02 09:55:19 crc kubenswrapper[5035]: I1002 09:55:19.062632 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-knpjm"] Oct 02 09:55:20 crc kubenswrapper[5035]: I1002 09:55:20.174819 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="136b4da6-5c57-4a29-af5a-afe32c5f552f" path="/var/lib/kubelet/pods/136b4da6-5c57-4a29-af5a-afe32c5f552f/volumes" Oct 02 09:55:20 crc kubenswrapper[5035]: I1002 09:55:20.176167 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ad8a14e-58b5-4192-acf8-972ffdbd8e27" path="/var/lib/kubelet/pods/2ad8a14e-58b5-4192-acf8-972ffdbd8e27/volumes" Oct 02 09:55:24 crc kubenswrapper[5035]: I1002 09:55:24.163225 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:55:24 crc kubenswrapper[5035]: E1002 09:55:24.163946 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:55:29 crc kubenswrapper[5035]: I1002 09:55:29.048773 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-d6j5s"] Oct 02 09:55:29 crc kubenswrapper[5035]: I1002 09:55:29.059737 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-d6j5s"] Oct 02 09:55:30 crc kubenswrapper[5035]: I1002 09:55:30.173460 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47413f27-08ae-470d-8877-84b1fd25fd8a" path="/var/lib/kubelet/pods/47413f27-08ae-470d-8877-84b1fd25fd8a/volumes" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.163340 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:55:36 crc kubenswrapper[5035]: E1002 09:55:36.164106 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.504723 5035 scope.go:117] "RemoveContainer" containerID="02b19eba82aa11ef3716fd2aa1d250359f0f75877f48cae54ce55304d1c6e091" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.546033 5035 scope.go:117] "RemoveContainer" containerID="d2218c6995a9867a45454a9cf044c0a69107d1783052cf86a94dcb9b1227b334" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.602144 5035 scope.go:117] "RemoveContainer" containerID="2db48747e2cf060ad12fdbfbc2f7b4c224d81ea2f5170ffee9e1d4692705084f" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.622824 5035 scope.go:117] "RemoveContainer" containerID="53ac0626fdd0ddba759654706b8203018a894f59965851289a2b8475b311e981" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.690463 5035 scope.go:117] "RemoveContainer" containerID="af2fc1b21d33c06faf5156a6bb204aed8e2faadf70b5f77609fdd6dd962f9909" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.719501 5035 scope.go:117] "RemoveContainer" containerID="fd0b46b7da0368c1ededacc7b6bbcccc49d456086e7701378108a883922d0c96" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.776742 5035 scope.go:117] "RemoveContainer" containerID="5c0780b6a2f1cfeefbb3baf3deb3f58cc4fbf0e9169e435c3d96f74f4e4393a9" Oct 02 09:55:36 crc kubenswrapper[5035]: I1002 09:55:36.797544 5035 scope.go:117] "RemoveContainer" containerID="16ad5aded9c7a6befe8479af76936081fa37f639cc6ade8422791922b3bf1318" Oct 02 09:55:37 crc kubenswrapper[5035]: I1002 09:55:37.031112 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-qf6x5"] Oct 02 09:55:37 crc kubenswrapper[5035]: I1002 09:55:37.043800 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-qf6x5"] Oct 02 09:55:38 crc kubenswrapper[5035]: I1002 09:55:38.179136 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="830b460b-be63-4b42-a442-80d844ef6908" path="/var/lib/kubelet/pods/830b460b-be63-4b42-a442-80d844ef6908/volumes" Oct 02 09:55:41 crc kubenswrapper[5035]: I1002 09:55:41.032621 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7t2tz"] Oct 02 09:55:41 crc kubenswrapper[5035]: I1002 09:55:41.044106 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7t2tz"] Oct 02 09:55:42 crc kubenswrapper[5035]: I1002 09:55:42.177113 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76ce8ea1-cda3-4f68-9570-c6c02b56283f" path="/var/lib/kubelet/pods/76ce8ea1-cda3-4f68-9570-c6c02b56283f/volumes" Oct 02 09:55:50 crc kubenswrapper[5035]: I1002 09:55:50.163410 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:55:50 crc kubenswrapper[5035]: E1002 09:55:50.164110 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:55:53 crc kubenswrapper[5035]: I1002 09:55:53.034369 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-db-create-l42l7"] Oct 02 09:55:53 crc kubenswrapper[5035]: I1002 09:55:53.041440 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-db-create-l42l7"] Oct 02 09:55:54 crc kubenswrapper[5035]: I1002 09:55:54.173938 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e575f2f-c857-4a4f-a771-e591df208a9c" path="/var/lib/kubelet/pods/7e575f2f-c857-4a4f-a771-e591df208a9c/volumes" Oct 02 09:56:03 crc kubenswrapper[5035]: I1002 09:56:03.163617 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:56:03 crc kubenswrapper[5035]: E1002 09:56:03.164343 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:56:13 crc kubenswrapper[5035]: I1002 09:56:13.033517 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-45c92"] Oct 02 09:56:13 crc kubenswrapper[5035]: I1002 09:56:13.043293 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-5076-account-create-hbnlx"] Oct 02 09:56:13 crc kubenswrapper[5035]: I1002 09:56:13.052927 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-45c92"] Oct 02 09:56:13 crc kubenswrapper[5035]: I1002 09:56:13.061603 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-5076-account-create-hbnlx"] Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.039204 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-6hhkr"] Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.047918 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-mp9fm"] Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.056258 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-mp9fm"] Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.064016 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-6hhkr"] Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.173735 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000c2597-bc6a-42d1-96e7-d622bbfe96f4" path="/var/lib/kubelet/pods/000c2597-bc6a-42d1-96e7-d622bbfe96f4/volumes" Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.174284 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a7815e7-389b-44fb-bcaf-feb2c60f8cb5" path="/var/lib/kubelet/pods/7a7815e7-389b-44fb-bcaf-feb2c60f8cb5/volumes" Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.174801 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="828d54cf-d9df-4c72-8010-3a39e3e986b7" path="/var/lib/kubelet/pods/828d54cf-d9df-4c72-8010-3a39e3e986b7/volumes" Oct 02 09:56:14 crc kubenswrapper[5035]: I1002 09:56:14.175255 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d" path="/var/lib/kubelet/pods/b1e2d42a-fb5a-4001-acac-0b5bbcb4b01d/volumes" Oct 02 09:56:18 crc kubenswrapper[5035]: I1002 09:56:18.163224 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:56:18 crc kubenswrapper[5035]: E1002 09:56:18.163976 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:56:29 crc kubenswrapper[5035]: I1002 09:56:29.041250 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-d8be-account-create-gxvzq"] Oct 02 09:56:29 crc kubenswrapper[5035]: I1002 09:56:29.049645 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-d8be-account-create-gxvzq"] Oct 02 09:56:30 crc kubenswrapper[5035]: I1002 09:56:30.197264 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43396b2-6fd0-4736-a8ba-796f27d0d461" path="/var/lib/kubelet/pods/f43396b2-6fd0-4736-a8ba-796f27d0d461/volumes" Oct 02 09:56:33 crc kubenswrapper[5035]: I1002 09:56:33.162951 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:56:33 crc kubenswrapper[5035]: E1002 09:56:33.163822 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:56:35 crc kubenswrapper[5035]: I1002 09:56:35.032052 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-1f39-account-create-2j8xx"] Oct 02 09:56:35 crc kubenswrapper[5035]: I1002 09:56:35.041345 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1475-account-create-7fvcf"] Oct 02 09:56:35 crc kubenswrapper[5035]: I1002 09:56:35.051393 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-1f39-account-create-2j8xx"] Oct 02 09:56:35 crc kubenswrapper[5035]: I1002 09:56:35.058272 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1475-account-create-7fvcf"] Oct 02 09:56:36 crc kubenswrapper[5035]: I1002 09:56:36.175326 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bee75d89-2175-4d59-bd8a-ff46df5eb123" path="/var/lib/kubelet/pods/bee75d89-2175-4d59-bd8a-ff46df5eb123/volumes" Oct 02 09:56:36 crc kubenswrapper[5035]: I1002 09:56:36.176297 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cacd2b45-3487-4e63-848d-466495eb599e" path="/var/lib/kubelet/pods/cacd2b45-3487-4e63-848d-466495eb599e/volumes" Oct 02 09:56:36 crc kubenswrapper[5035]: I1002 09:56:36.950641 5035 scope.go:117] "RemoveContainer" containerID="6a4f10c5814a8ae7e34611ca2236badbceaa7dd1b30aa7238bd4b1a43e72fa80" Oct 02 09:56:36 crc kubenswrapper[5035]: I1002 09:56:36.990104 5035 scope.go:117] "RemoveContainer" containerID="81ffd6e467018eb41bdb851c2938aa76b37f33fe707a1335da83fa766cc6f684" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.054930 5035 scope.go:117] "RemoveContainer" containerID="d7010f1e2875d85efddf261d1615f7a7534deda7f841e7995e0a969330ce9058" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.079708 5035 scope.go:117] "RemoveContainer" containerID="4b7d1a789544379cb5e1fd57aed70c0a43a1d8fec4ddd71aa06c6f6e7989b772" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.121636 5035 scope.go:117] "RemoveContainer" containerID="a3c54c4b74f2501aebce1fdc725b64e3df3f6281017bdbc507ead15479f2b2b5" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.171060 5035 scope.go:117] "RemoveContainer" containerID="ecc30c7d8b0724e3898e58106086877159481cd668ab0423fe8da23f1a477b6e" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.210925 5035 scope.go:117] "RemoveContainer" containerID="c2429c615c86013bad391a92de956173d184172837bf5ba13809e7e863988ef4" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.229590 5035 scope.go:117] "RemoveContainer" containerID="49e6bc1dcd3210e3ae2c9a69fafc0c053a9f33d4a050ddbc3b8504564fdd8373" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.264053 5035 scope.go:117] "RemoveContainer" containerID="2b9b61429e301eff6bedaadffba119544db4e63fa18686a2c34f5143af4c0b0e" Oct 02 09:56:37 crc kubenswrapper[5035]: I1002 09:56:37.283353 5035 scope.go:117] "RemoveContainer" containerID="108fbb5286e8dabf702a656539736f6cb58c43a311227b9c5b2d0933303d0e57" Oct 02 09:56:47 crc kubenswrapper[5035]: I1002 09:56:47.163227 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:56:47 crc kubenswrapper[5035]: E1002 09:56:47.163947 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:57:02 crc kubenswrapper[5035]: I1002 09:57:02.162953 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:57:02 crc kubenswrapper[5035]: E1002 09:57:02.163772 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:57:15 crc kubenswrapper[5035]: I1002 09:57:15.163043 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:57:15 crc kubenswrapper[5035]: E1002 09:57:15.163982 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:57:29 crc kubenswrapper[5035]: I1002 09:57:29.163665 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:57:29 crc kubenswrapper[5035]: E1002 09:57:29.164323 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:57:33 crc kubenswrapper[5035]: I1002 09:57:33.040255 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sgtqw"] Oct 02 09:57:33 crc kubenswrapper[5035]: I1002 09:57:33.049086 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sgtqw"] Oct 02 09:57:34 crc kubenswrapper[5035]: I1002 09:57:34.173900 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ae7cc3-99fc-4485-b5b3-589d64791997" path="/var/lib/kubelet/pods/54ae7cc3-99fc-4485-b5b3-589d64791997/volumes" Oct 02 09:57:37 crc kubenswrapper[5035]: I1002 09:57:37.453417 5035 scope.go:117] "RemoveContainer" containerID="7f617c73dfbdcae4fe1152dd876e3ad4da1920b515ac23133acb5a9b3270619c" Oct 02 09:57:42 crc kubenswrapper[5035]: I1002 09:57:42.164474 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:57:42 crc kubenswrapper[5035]: E1002 09:57:42.165944 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:57:56 crc kubenswrapper[5035]: I1002 09:57:56.053375 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-7jwqw"] Oct 02 09:57:56 crc kubenswrapper[5035]: I1002 09:57:56.061149 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-7jwqw"] Oct 02 09:57:56 crc kubenswrapper[5035]: I1002 09:57:56.174166 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="104b87a6-20cd-459e-b095-eaa7de00ba31" path="/var/lib/kubelet/pods/104b87a6-20cd-459e-b095-eaa7de00ba31/volumes" Oct 02 09:57:57 crc kubenswrapper[5035]: I1002 09:57:57.164454 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:57:57 crc kubenswrapper[5035]: E1002 09:57:57.164759 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:58:00 crc kubenswrapper[5035]: I1002 09:58:00.031763 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z7gpk"] Oct 02 09:58:00 crc kubenswrapper[5035]: I1002 09:58:00.040427 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-z7gpk"] Oct 02 09:58:00 crc kubenswrapper[5035]: I1002 09:58:00.178370 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55b701fb-25b6-4e11-b46c-08a523c0d0ea" path="/var/lib/kubelet/pods/55b701fb-25b6-4e11-b46c-08a523c0d0ea/volumes" Oct 02 09:58:10 crc kubenswrapper[5035]: I1002 09:58:10.166270 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:58:10 crc kubenswrapper[5035]: E1002 09:58:10.173636 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:58:23 crc kubenswrapper[5035]: I1002 09:58:23.163041 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:58:23 crc kubenswrapper[5035]: E1002 09:58:23.164252 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 09:58:37 crc kubenswrapper[5035]: I1002 09:58:37.541515 5035 scope.go:117] "RemoveContainer" containerID="083b2ea6a1503b549b21bb73e43ad1feecf7118e404dfcfbed8a661ed4df1bc4" Oct 02 09:58:37 crc kubenswrapper[5035]: I1002 09:58:37.587384 5035 scope.go:117] "RemoveContainer" containerID="e73d2ddf7b019bd706f80ab956f457a6e25470d37e54168a6c1f392c093fd74f" Oct 02 09:58:38 crc kubenswrapper[5035]: I1002 09:58:38.163212 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 09:58:39 crc kubenswrapper[5035]: I1002 09:58:39.392854 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"b79bebf9853f6791ba16c2df809895df6aa63570391981a1ed44263bb7d7dbe5"} Oct 02 09:58:41 crc kubenswrapper[5035]: I1002 09:58:41.041229 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsl64"] Oct 02 09:58:41 crc kubenswrapper[5035]: I1002 09:58:41.048136 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dsl64"] Oct 02 09:58:42 crc kubenswrapper[5035]: I1002 09:58:42.173796 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093232c1-8151-4eac-97ec-76c29f228e05" path="/var/lib/kubelet/pods/093232c1-8151-4eac-97ec-76c29f228e05/volumes" Oct 02 09:59:37 crc kubenswrapper[5035]: I1002 09:59:37.679653 5035 scope.go:117] "RemoveContainer" containerID="712e3a0139796a4a1434856e06363d8133baf9310624eb20d5916f30f3d4fd86" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.149394 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq"] Oct 02 10:00:00 crc kubenswrapper[5035]: E1002 10:00:00.150439 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="extract-utilities" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.150454 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="extract-utilities" Oct 02 10:00:00 crc kubenswrapper[5035]: E1002 10:00:00.150463 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="registry-server" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.150471 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="registry-server" Oct 02 10:00:00 crc kubenswrapper[5035]: E1002 10:00:00.150493 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="extract-content" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.150500 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="extract-content" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.150747 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="54958464-af3c-45c5-b4bd-bc1a07f6f758" containerName="registry-server" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.152581 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.159585 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.159712 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.161693 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq"] Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.341468 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svlfb\" (UniqueName: \"kubernetes.io/projected/e63b90f7-8f1c-4d23-b690-33e87171fe86-kube-api-access-svlfb\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.341642 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e63b90f7-8f1c-4d23-b690-33e87171fe86-config-volume\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.341716 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e63b90f7-8f1c-4d23-b690-33e87171fe86-secret-volume\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.443640 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e63b90f7-8f1c-4d23-b690-33e87171fe86-config-volume\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.443737 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e63b90f7-8f1c-4d23-b690-33e87171fe86-secret-volume\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.443819 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svlfb\" (UniqueName: \"kubernetes.io/projected/e63b90f7-8f1c-4d23-b690-33e87171fe86-kube-api-access-svlfb\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.444782 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e63b90f7-8f1c-4d23-b690-33e87171fe86-config-volume\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.451298 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e63b90f7-8f1c-4d23-b690-33e87171fe86-secret-volume\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.464775 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svlfb\" (UniqueName: \"kubernetes.io/projected/e63b90f7-8f1c-4d23-b690-33e87171fe86-kube-api-access-svlfb\") pod \"collect-profiles-29323320-rcjbq\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.478960 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:00 crc kubenswrapper[5035]: I1002 10:00:00.918140 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq"] Oct 02 10:00:01 crc kubenswrapper[5035]: I1002 10:00:01.089579 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" event={"ID":"e63b90f7-8f1c-4d23-b690-33e87171fe86","Type":"ContainerStarted","Data":"ace1522630809e2c6f43a47eb8faa0054441f2050485246148e26c4178a3a623"} Oct 02 10:00:02 crc kubenswrapper[5035]: I1002 10:00:02.100002 5035 generic.go:334] "Generic (PLEG): container finished" podID="e63b90f7-8f1c-4d23-b690-33e87171fe86" containerID="1a74d5d5161c9c6e0d9d4cb7ab8606fffc59df0c53db174c8a136be63a0deaa6" exitCode=0 Oct 02 10:00:02 crc kubenswrapper[5035]: I1002 10:00:02.100100 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" event={"ID":"e63b90f7-8f1c-4d23-b690-33e87171fe86","Type":"ContainerDied","Data":"1a74d5d5161c9c6e0d9d4cb7ab8606fffc59df0c53db174c8a136be63a0deaa6"} Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.411601 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.500626 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e63b90f7-8f1c-4d23-b690-33e87171fe86-config-volume\") pod \"e63b90f7-8f1c-4d23-b690-33e87171fe86\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.500685 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svlfb\" (UniqueName: \"kubernetes.io/projected/e63b90f7-8f1c-4d23-b690-33e87171fe86-kube-api-access-svlfb\") pod \"e63b90f7-8f1c-4d23-b690-33e87171fe86\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.500749 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e63b90f7-8f1c-4d23-b690-33e87171fe86-secret-volume\") pod \"e63b90f7-8f1c-4d23-b690-33e87171fe86\" (UID: \"e63b90f7-8f1c-4d23-b690-33e87171fe86\") " Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.501330 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e63b90f7-8f1c-4d23-b690-33e87171fe86-config-volume" (OuterVolumeSpecName: "config-volume") pod "e63b90f7-8f1c-4d23-b690-33e87171fe86" (UID: "e63b90f7-8f1c-4d23-b690-33e87171fe86"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.506096 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e63b90f7-8f1c-4d23-b690-33e87171fe86-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e63b90f7-8f1c-4d23-b690-33e87171fe86" (UID: "e63b90f7-8f1c-4d23-b690-33e87171fe86"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.507149 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e63b90f7-8f1c-4d23-b690-33e87171fe86-kube-api-access-svlfb" (OuterVolumeSpecName: "kube-api-access-svlfb") pod "e63b90f7-8f1c-4d23-b690-33e87171fe86" (UID: "e63b90f7-8f1c-4d23-b690-33e87171fe86"). InnerVolumeSpecName "kube-api-access-svlfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.602926 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e63b90f7-8f1c-4d23-b690-33e87171fe86-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.602956 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svlfb\" (UniqueName: \"kubernetes.io/projected/e63b90f7-8f1c-4d23-b690-33e87171fe86-kube-api-access-svlfb\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:03 crc kubenswrapper[5035]: I1002 10:00:03.602967 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e63b90f7-8f1c-4d23-b690-33e87171fe86-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:00:04 crc kubenswrapper[5035]: I1002 10:00:04.118920 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" event={"ID":"e63b90f7-8f1c-4d23-b690-33e87171fe86","Type":"ContainerDied","Data":"ace1522630809e2c6f43a47eb8faa0054441f2050485246148e26c4178a3a623"} Oct 02 10:00:04 crc kubenswrapper[5035]: I1002 10:00:04.119273 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ace1522630809e2c6f43a47eb8faa0054441f2050485246148e26c4178a3a623" Oct 02 10:00:04 crc kubenswrapper[5035]: I1002 10:00:04.119076 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323320-rcjbq" Oct 02 10:00:55 crc kubenswrapper[5035]: I1002 10:00:55.538518 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:00:55 crc kubenswrapper[5035]: I1002 10:00:55.539661 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.154521 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323321-qsw95"] Oct 02 10:01:00 crc kubenswrapper[5035]: E1002 10:01:00.156919 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63b90f7-8f1c-4d23-b690-33e87171fe86" containerName="collect-profiles" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.156953 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63b90f7-8f1c-4d23-b690-33e87171fe86" containerName="collect-profiles" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.157227 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63b90f7-8f1c-4d23-b690-33e87171fe86" containerName="collect-profiles" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.158029 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.179961 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323321-qsw95"] Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.296175 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-config-data\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.296614 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-fernet-keys\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.296674 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjg8\" (UniqueName: \"kubernetes.io/projected/1229f682-2732-4b72-aa97-2443db2ebf64-kube-api-access-vkjg8\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.296698 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-combined-ca-bundle\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.398115 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-fernet-keys\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.398218 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjg8\" (UniqueName: \"kubernetes.io/projected/1229f682-2732-4b72-aa97-2443db2ebf64-kube-api-access-vkjg8\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.398247 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-combined-ca-bundle\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.398414 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-config-data\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.405218 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-config-data\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.405321 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-fernet-keys\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.412978 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-combined-ca-bundle\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.415641 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjg8\" (UniqueName: \"kubernetes.io/projected/1229f682-2732-4b72-aa97-2443db2ebf64-kube-api-access-vkjg8\") pod \"keystone-cron-29323321-qsw95\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.494254 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:00 crc kubenswrapper[5035]: I1002 10:01:00.926808 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323321-qsw95"] Oct 02 10:01:00 crc kubenswrapper[5035]: W1002 10:01:00.929360 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1229f682_2732_4b72_aa97_2443db2ebf64.slice/crio-9aa2a6948f86cd8a440431fadeec108d0bb8167c61e7393207d7d069f3545c24 WatchSource:0}: Error finding container 9aa2a6948f86cd8a440431fadeec108d0bb8167c61e7393207d7d069f3545c24: Status 404 returned error can't find the container with id 9aa2a6948f86cd8a440431fadeec108d0bb8167c61e7393207d7d069f3545c24 Oct 02 10:01:01 crc kubenswrapper[5035]: I1002 10:01:01.633508 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-qsw95" event={"ID":"1229f682-2732-4b72-aa97-2443db2ebf64","Type":"ContainerStarted","Data":"089aec50347722f7abbd7cebbc6a53d2d3cc983b9a2e864da52793045fb8d316"} Oct 02 10:01:01 crc kubenswrapper[5035]: I1002 10:01:01.633857 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-qsw95" event={"ID":"1229f682-2732-4b72-aa97-2443db2ebf64","Type":"ContainerStarted","Data":"9aa2a6948f86cd8a440431fadeec108d0bb8167c61e7393207d7d069f3545c24"} Oct 02 10:01:01 crc kubenswrapper[5035]: I1002 10:01:01.654691 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323321-qsw95" podStartSLOduration=1.654670925 podStartE2EDuration="1.654670925s" podCreationTimestamp="2025-10-02 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:01:01.648583597 +0000 UTC m=+2027.004927642" watchObservedRunningTime="2025-10-02 10:01:01.654670925 +0000 UTC m=+2027.011014960" Oct 02 10:01:03 crc kubenswrapper[5035]: I1002 10:01:03.651175 5035 generic.go:334] "Generic (PLEG): container finished" podID="1229f682-2732-4b72-aa97-2443db2ebf64" containerID="089aec50347722f7abbd7cebbc6a53d2d3cc983b9a2e864da52793045fb8d316" exitCode=0 Oct 02 10:01:03 crc kubenswrapper[5035]: I1002 10:01:03.651244 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-qsw95" event={"ID":"1229f682-2732-4b72-aa97-2443db2ebf64","Type":"ContainerDied","Data":"089aec50347722f7abbd7cebbc6a53d2d3cc983b9a2e864da52793045fb8d316"} Oct 02 10:01:04 crc kubenswrapper[5035]: I1002 10:01:04.987390 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.087795 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-config-data\") pod \"1229f682-2732-4b72-aa97-2443db2ebf64\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.087940 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkjg8\" (UniqueName: \"kubernetes.io/projected/1229f682-2732-4b72-aa97-2443db2ebf64-kube-api-access-vkjg8\") pod \"1229f682-2732-4b72-aa97-2443db2ebf64\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.088688 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-fernet-keys\") pod \"1229f682-2732-4b72-aa97-2443db2ebf64\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.088865 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-combined-ca-bundle\") pod \"1229f682-2732-4b72-aa97-2443db2ebf64\" (UID: \"1229f682-2732-4b72-aa97-2443db2ebf64\") " Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.094821 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1229f682-2732-4b72-aa97-2443db2ebf64" (UID: "1229f682-2732-4b72-aa97-2443db2ebf64"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.095212 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1229f682-2732-4b72-aa97-2443db2ebf64-kube-api-access-vkjg8" (OuterVolumeSpecName: "kube-api-access-vkjg8") pod "1229f682-2732-4b72-aa97-2443db2ebf64" (UID: "1229f682-2732-4b72-aa97-2443db2ebf64"). InnerVolumeSpecName "kube-api-access-vkjg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.122685 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1229f682-2732-4b72-aa97-2443db2ebf64" (UID: "1229f682-2732-4b72-aa97-2443db2ebf64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.147669 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-config-data" (OuterVolumeSpecName: "config-data") pod "1229f682-2732-4b72-aa97-2443db2ebf64" (UID: "1229f682-2732-4b72-aa97-2443db2ebf64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.191828 5035 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.191869 5035 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.191879 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkjg8\" (UniqueName: \"kubernetes.io/projected/1229f682-2732-4b72-aa97-2443db2ebf64-kube-api-access-vkjg8\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.191889 5035 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1229f682-2732-4b72-aa97-2443db2ebf64-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.674674 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323321-qsw95" event={"ID":"1229f682-2732-4b72-aa97-2443db2ebf64","Type":"ContainerDied","Data":"9aa2a6948f86cd8a440431fadeec108d0bb8167c61e7393207d7d069f3545c24"} Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.674708 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9aa2a6948f86cd8a440431fadeec108d0bb8167c61e7393207d7d069f3545c24" Oct 02 10:01:05 crc kubenswrapper[5035]: I1002 10:01:05.674744 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323321-qsw95" Oct 02 10:01:25 crc kubenswrapper[5035]: I1002 10:01:25.538068 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:01:25 crc kubenswrapper[5035]: I1002 10:01:25.538675 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:01:55 crc kubenswrapper[5035]: I1002 10:01:55.537566 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:01:55 crc kubenswrapper[5035]: I1002 10:01:55.538819 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:01:55 crc kubenswrapper[5035]: I1002 10:01:55.538894 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 10:01:55 crc kubenswrapper[5035]: I1002 10:01:55.539719 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b79bebf9853f6791ba16c2df809895df6aa63570391981a1ed44263bb7d7dbe5"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:01:55 crc kubenswrapper[5035]: I1002 10:01:55.539799 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://b79bebf9853f6791ba16c2df809895df6aa63570391981a1ed44263bb7d7dbe5" gracePeriod=600 Oct 02 10:01:56 crc kubenswrapper[5035]: I1002 10:01:56.102198 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="b79bebf9853f6791ba16c2df809895df6aa63570391981a1ed44263bb7d7dbe5" exitCode=0 Oct 02 10:01:56 crc kubenswrapper[5035]: I1002 10:01:56.102292 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"b79bebf9853f6791ba16c2df809895df6aa63570391981a1ed44263bb7d7dbe5"} Oct 02 10:01:56 crc kubenswrapper[5035]: I1002 10:01:56.102637 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9"} Oct 02 10:01:56 crc kubenswrapper[5035]: I1002 10:01:56.102666 5035 scope.go:117] "RemoveContainer" containerID="d97c6807749aed40109ff3a7d279ce92a06b772f92c2ea53c5690b6b5e1d2363" Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.875790 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-m7gcv"] Oct 02 10:02:24 crc kubenswrapper[5035]: E1002 10:02:24.876893 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1229f682-2732-4b72-aa97-2443db2ebf64" containerName="keystone-cron" Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.876912 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1229f682-2732-4b72-aa97-2443db2ebf64" containerName="keystone-cron" Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.877146 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1229f682-2732-4b72-aa97-2443db2ebf64" containerName="keystone-cron" Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.878997 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.890268 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m7gcv"] Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.936615 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-catalog-content\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.936661 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-utilities\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:24 crc kubenswrapper[5035]: I1002 10:02:24.936782 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ffhh\" (UniqueName: \"kubernetes.io/projected/89bf060a-9f64-4c15-a2f1-f3de264503c9-kube-api-access-9ffhh\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.038486 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-catalog-content\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.038555 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-utilities\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.038622 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ffhh\" (UniqueName: \"kubernetes.io/projected/89bf060a-9f64-4c15-a2f1-f3de264503c9-kube-api-access-9ffhh\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.039046 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-catalog-content\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.039206 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-utilities\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.056871 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ffhh\" (UniqueName: \"kubernetes.io/projected/89bf060a-9f64-4c15-a2f1-f3de264503c9-kube-api-access-9ffhh\") pod \"redhat-operators-m7gcv\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.203652 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:25 crc kubenswrapper[5035]: I1002 10:02:25.664291 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-m7gcv"] Oct 02 10:02:26 crc kubenswrapper[5035]: I1002 10:02:26.358316 5035 generic.go:334] "Generic (PLEG): container finished" podID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerID="607ed028ae33cc35d894e93851c633b728d0136b22df00f62e344973ac900a5d" exitCode=0 Oct 02 10:02:26 crc kubenswrapper[5035]: I1002 10:02:26.358381 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7gcv" event={"ID":"89bf060a-9f64-4c15-a2f1-f3de264503c9","Type":"ContainerDied","Data":"607ed028ae33cc35d894e93851c633b728d0136b22df00f62e344973ac900a5d"} Oct 02 10:02:26 crc kubenswrapper[5035]: I1002 10:02:26.358621 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7gcv" event={"ID":"89bf060a-9f64-4c15-a2f1-f3de264503c9","Type":"ContainerStarted","Data":"0262272f91472f70e0076b3605a1eb6fc44f28ca2018fb169d2f7f084c72024f"} Oct 02 10:02:26 crc kubenswrapper[5035]: I1002 10:02:26.361416 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:02:27 crc kubenswrapper[5035]: I1002 10:02:27.368958 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7gcv" event={"ID":"89bf060a-9f64-4c15-a2f1-f3de264503c9","Type":"ContainerStarted","Data":"53318a07ff47f101aa76db2eb55f4c9bca29a939c188e731c373472dbdacf728"} Oct 02 10:02:28 crc kubenswrapper[5035]: I1002 10:02:28.383044 5035 generic.go:334] "Generic (PLEG): container finished" podID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerID="53318a07ff47f101aa76db2eb55f4c9bca29a939c188e731c373472dbdacf728" exitCode=0 Oct 02 10:02:28 crc kubenswrapper[5035]: I1002 10:02:28.383184 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7gcv" event={"ID":"89bf060a-9f64-4c15-a2f1-f3de264503c9","Type":"ContainerDied","Data":"53318a07ff47f101aa76db2eb55f4c9bca29a939c188e731c373472dbdacf728"} Oct 02 10:02:29 crc kubenswrapper[5035]: I1002 10:02:29.398906 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7gcv" event={"ID":"89bf060a-9f64-4c15-a2f1-f3de264503c9","Type":"ContainerStarted","Data":"bf01569fad94594e94018bfb080796a119e21ed3d8b9e9fc547cee52013b8ffd"} Oct 02 10:02:35 crc kubenswrapper[5035]: I1002 10:02:35.204745 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:35 crc kubenswrapper[5035]: I1002 10:02:35.206674 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:35 crc kubenswrapper[5035]: I1002 10:02:35.254244 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:35 crc kubenswrapper[5035]: I1002 10:02:35.273315 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-m7gcv" podStartSLOduration=8.714322145 podStartE2EDuration="11.273294232s" podCreationTimestamp="2025-10-02 10:02:24 +0000 UTC" firstStartedPulling="2025-10-02 10:02:26.361100038 +0000 UTC m=+2111.717444063" lastFinishedPulling="2025-10-02 10:02:28.920072125 +0000 UTC m=+2114.276416150" observedRunningTime="2025-10-02 10:02:29.421703128 +0000 UTC m=+2114.778047153" watchObservedRunningTime="2025-10-02 10:02:35.273294232 +0000 UTC m=+2120.629638257" Oct 02 10:02:35 crc kubenswrapper[5035]: I1002 10:02:35.517990 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:35 crc kubenswrapper[5035]: I1002 10:02:35.570076 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m7gcv"] Oct 02 10:02:37 crc kubenswrapper[5035]: I1002 10:02:37.487315 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-m7gcv" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="registry-server" containerID="cri-o://bf01569fad94594e94018bfb080796a119e21ed3d8b9e9fc547cee52013b8ffd" gracePeriod=2 Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.504235 5035 generic.go:334] "Generic (PLEG): container finished" podID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerID="bf01569fad94594e94018bfb080796a119e21ed3d8b9e9fc547cee52013b8ffd" exitCode=0 Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.505720 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7gcv" event={"ID":"89bf060a-9f64-4c15-a2f1-f3de264503c9","Type":"ContainerDied","Data":"bf01569fad94594e94018bfb080796a119e21ed3d8b9e9fc547cee52013b8ffd"} Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.794788 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.911313 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-utilities\") pod \"89bf060a-9f64-4c15-a2f1-f3de264503c9\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.911406 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-catalog-content\") pod \"89bf060a-9f64-4c15-a2f1-f3de264503c9\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.911471 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ffhh\" (UniqueName: \"kubernetes.io/projected/89bf060a-9f64-4c15-a2f1-f3de264503c9-kube-api-access-9ffhh\") pod \"89bf060a-9f64-4c15-a2f1-f3de264503c9\" (UID: \"89bf060a-9f64-4c15-a2f1-f3de264503c9\") " Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.912185 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-utilities" (OuterVolumeSpecName: "utilities") pod "89bf060a-9f64-4c15-a2f1-f3de264503c9" (UID: "89bf060a-9f64-4c15-a2f1-f3de264503c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.917609 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89bf060a-9f64-4c15-a2f1-f3de264503c9-kube-api-access-9ffhh" (OuterVolumeSpecName: "kube-api-access-9ffhh") pod "89bf060a-9f64-4c15-a2f1-f3de264503c9" (UID: "89bf060a-9f64-4c15-a2f1-f3de264503c9"). InnerVolumeSpecName "kube-api-access-9ffhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:02:38 crc kubenswrapper[5035]: I1002 10:02:38.988306 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89bf060a-9f64-4c15-a2f1-f3de264503c9" (UID: "89bf060a-9f64-4c15-a2f1-f3de264503c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.013054 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.013090 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89bf060a-9f64-4c15-a2f1-f3de264503c9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.013105 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ffhh\" (UniqueName: \"kubernetes.io/projected/89bf060a-9f64-4c15-a2f1-f3de264503c9-kube-api-access-9ffhh\") on node \"crc\" DevicePath \"\"" Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.515678 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-m7gcv" event={"ID":"89bf060a-9f64-4c15-a2f1-f3de264503c9","Type":"ContainerDied","Data":"0262272f91472f70e0076b3605a1eb6fc44f28ca2018fb169d2f7f084c72024f"} Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.516128 5035 scope.go:117] "RemoveContainer" containerID="bf01569fad94594e94018bfb080796a119e21ed3d8b9e9fc547cee52013b8ffd" Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.515741 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-m7gcv" Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.540847 5035 scope.go:117] "RemoveContainer" containerID="53318a07ff47f101aa76db2eb55f4c9bca29a939c188e731c373472dbdacf728" Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.562679 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-m7gcv"] Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.568215 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-m7gcv"] Oct 02 10:02:39 crc kubenswrapper[5035]: I1002 10:02:39.587016 5035 scope.go:117] "RemoveContainer" containerID="607ed028ae33cc35d894e93851c633b728d0136b22df00f62e344973ac900a5d" Oct 02 10:02:40 crc kubenswrapper[5035]: I1002 10:02:40.174628 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" path="/var/lib/kubelet/pods/89bf060a-9f64-4c15-a2f1-f3de264503c9/volumes" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.249251 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rqrgr"] Oct 02 10:03:08 crc kubenswrapper[5035]: E1002 10:03:08.250289 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="extract-utilities" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.250304 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="extract-utilities" Oct 02 10:03:08 crc kubenswrapper[5035]: E1002 10:03:08.250338 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="extract-content" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.250346 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="extract-content" Oct 02 10:03:08 crc kubenswrapper[5035]: E1002 10:03:08.250382 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="registry-server" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.250391 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="registry-server" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.250634 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="89bf060a-9f64-4c15-a2f1-f3de264503c9" containerName="registry-server" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.252306 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.268077 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-catalog-content\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.268254 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgj5g\" (UniqueName: \"kubernetes.io/projected/bf95e53c-b31a-4118-a403-5e8764bf0aa3-kube-api-access-fgj5g\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.268321 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-utilities\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.275184 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqrgr"] Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.370059 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-catalog-content\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.370150 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgj5g\" (UniqueName: \"kubernetes.io/projected/bf95e53c-b31a-4118-a403-5e8764bf0aa3-kube-api-access-fgj5g\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.370190 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-utilities\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.370789 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-utilities\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.370832 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-catalog-content\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.395897 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgj5g\" (UniqueName: \"kubernetes.io/projected/bf95e53c-b31a-4118-a403-5e8764bf0aa3-kube-api-access-fgj5g\") pod \"redhat-marketplace-rqrgr\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:08 crc kubenswrapper[5035]: I1002 10:03:08.578583 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:09 crc kubenswrapper[5035]: I1002 10:03:09.036241 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqrgr"] Oct 02 10:03:09 crc kubenswrapper[5035]: I1002 10:03:09.776613 5035 generic.go:334] "Generic (PLEG): container finished" podID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerID="aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3" exitCode=0 Oct 02 10:03:09 crc kubenswrapper[5035]: I1002 10:03:09.776680 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqrgr" event={"ID":"bf95e53c-b31a-4118-a403-5e8764bf0aa3","Type":"ContainerDied","Data":"aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3"} Oct 02 10:03:09 crc kubenswrapper[5035]: I1002 10:03:09.776962 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqrgr" event={"ID":"bf95e53c-b31a-4118-a403-5e8764bf0aa3","Type":"ContainerStarted","Data":"5bc1ab405ed91258619415d396d70dd09d3c21429dd750e2eb43cca6e2184127"} Oct 02 10:03:10 crc kubenswrapper[5035]: I1002 10:03:10.786887 5035 generic.go:334] "Generic (PLEG): container finished" podID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerID="b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324" exitCode=0 Oct 02 10:03:10 crc kubenswrapper[5035]: I1002 10:03:10.786943 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqrgr" event={"ID":"bf95e53c-b31a-4118-a403-5e8764bf0aa3","Type":"ContainerDied","Data":"b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324"} Oct 02 10:03:11 crc kubenswrapper[5035]: I1002 10:03:11.809801 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqrgr" event={"ID":"bf95e53c-b31a-4118-a403-5e8764bf0aa3","Type":"ContainerStarted","Data":"2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50"} Oct 02 10:03:11 crc kubenswrapper[5035]: I1002 10:03:11.830304 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rqrgr" podStartSLOduration=2.4011149 podStartE2EDuration="3.830286065s" podCreationTimestamp="2025-10-02 10:03:08 +0000 UTC" firstStartedPulling="2025-10-02 10:03:09.778640796 +0000 UTC m=+2155.134984831" lastFinishedPulling="2025-10-02 10:03:11.207811971 +0000 UTC m=+2156.564155996" observedRunningTime="2025-10-02 10:03:11.828817682 +0000 UTC m=+2157.185161707" watchObservedRunningTime="2025-10-02 10:03:11.830286065 +0000 UTC m=+2157.186630090" Oct 02 10:03:18 crc kubenswrapper[5035]: I1002 10:03:18.579760 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:18 crc kubenswrapper[5035]: I1002 10:03:18.580264 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:18 crc kubenswrapper[5035]: I1002 10:03:18.624900 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:18 crc kubenswrapper[5035]: I1002 10:03:18.913456 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:18 crc kubenswrapper[5035]: I1002 10:03:18.955816 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqrgr"] Oct 02 10:03:20 crc kubenswrapper[5035]: I1002 10:03:20.889226 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rqrgr" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="registry-server" containerID="cri-o://2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50" gracePeriod=2 Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.319033 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.421875 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-catalog-content\") pod \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.422293 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgj5g\" (UniqueName: \"kubernetes.io/projected/bf95e53c-b31a-4118-a403-5e8764bf0aa3-kube-api-access-fgj5g\") pod \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.422403 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-utilities\") pod \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\" (UID: \"bf95e53c-b31a-4118-a403-5e8764bf0aa3\") " Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.423570 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-utilities" (OuterVolumeSpecName: "utilities") pod "bf95e53c-b31a-4118-a403-5e8764bf0aa3" (UID: "bf95e53c-b31a-4118-a403-5e8764bf0aa3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.428988 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf95e53c-b31a-4118-a403-5e8764bf0aa3-kube-api-access-fgj5g" (OuterVolumeSpecName: "kube-api-access-fgj5g") pod "bf95e53c-b31a-4118-a403-5e8764bf0aa3" (UID: "bf95e53c-b31a-4118-a403-5e8764bf0aa3"). InnerVolumeSpecName "kube-api-access-fgj5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.435773 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf95e53c-b31a-4118-a403-5e8764bf0aa3" (UID: "bf95e53c-b31a-4118-a403-5e8764bf0aa3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.524206 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgj5g\" (UniqueName: \"kubernetes.io/projected/bf95e53c-b31a-4118-a403-5e8764bf0aa3-kube-api-access-fgj5g\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.524237 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.524246 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf95e53c-b31a-4118-a403-5e8764bf0aa3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.897778 5035 generic.go:334] "Generic (PLEG): container finished" podID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerID="2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50" exitCode=0 Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.897840 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqrgr" event={"ID":"bf95e53c-b31a-4118-a403-5e8764bf0aa3","Type":"ContainerDied","Data":"2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50"} Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.897846 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rqrgr" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.897891 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rqrgr" event={"ID":"bf95e53c-b31a-4118-a403-5e8764bf0aa3","Type":"ContainerDied","Data":"5bc1ab405ed91258619415d396d70dd09d3c21429dd750e2eb43cca6e2184127"} Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.897920 5035 scope.go:117] "RemoveContainer" containerID="2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.927606 5035 scope.go:117] "RemoveContainer" containerID="b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324" Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.939948 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqrgr"] Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.946872 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rqrgr"] Oct 02 10:03:21 crc kubenswrapper[5035]: I1002 10:03:21.965009 5035 scope.go:117] "RemoveContainer" containerID="aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3" Oct 02 10:03:22 crc kubenswrapper[5035]: I1002 10:03:22.002460 5035 scope.go:117] "RemoveContainer" containerID="2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50" Oct 02 10:03:22 crc kubenswrapper[5035]: E1002 10:03:22.003002 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50\": container with ID starting with 2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50 not found: ID does not exist" containerID="2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50" Oct 02 10:03:22 crc kubenswrapper[5035]: I1002 10:03:22.003053 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50"} err="failed to get container status \"2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50\": rpc error: code = NotFound desc = could not find container \"2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50\": container with ID starting with 2f42bd6f04c318e8e5da79e3a039f0a24fcc81f6b731b0c887adfb3101e9dc50 not found: ID does not exist" Oct 02 10:03:22 crc kubenswrapper[5035]: I1002 10:03:22.003074 5035 scope.go:117] "RemoveContainer" containerID="b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324" Oct 02 10:03:22 crc kubenswrapper[5035]: E1002 10:03:22.003353 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324\": container with ID starting with b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324 not found: ID does not exist" containerID="b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324" Oct 02 10:03:22 crc kubenswrapper[5035]: I1002 10:03:22.003384 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324"} err="failed to get container status \"b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324\": rpc error: code = NotFound desc = could not find container \"b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324\": container with ID starting with b77dd5068329906fc1ba5acb8d3d707d74593c9bd2bf00342aaa9ba49aeab324 not found: ID does not exist" Oct 02 10:03:22 crc kubenswrapper[5035]: I1002 10:03:22.003405 5035 scope.go:117] "RemoveContainer" containerID="aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3" Oct 02 10:03:22 crc kubenswrapper[5035]: E1002 10:03:22.003690 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3\": container with ID starting with aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3 not found: ID does not exist" containerID="aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3" Oct 02 10:03:22 crc kubenswrapper[5035]: I1002 10:03:22.003715 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3"} err="failed to get container status \"aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3\": rpc error: code = NotFound desc = could not find container \"aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3\": container with ID starting with aa940a5b340fed1bbec4b09ebc43fca249346e9ccee22c1480787fc418666ed3 not found: ID does not exist" Oct 02 10:03:22 crc kubenswrapper[5035]: I1002 10:03:22.174043 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" path="/var/lib/kubelet/pods/bf95e53c-b31a-4118-a403-5e8764bf0aa3/volumes" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.791445 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gr5ck"] Oct 02 10:03:46 crc kubenswrapper[5035]: E1002 10:03:46.793629 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="registry-server" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.793858 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="registry-server" Oct 02 10:03:46 crc kubenswrapper[5035]: E1002 10:03:46.793969 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="extract-utilities" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.794041 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="extract-utilities" Oct 02 10:03:46 crc kubenswrapper[5035]: E1002 10:03:46.794142 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="extract-content" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.794218 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="extract-content" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.794549 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf95e53c-b31a-4118-a403-5e8764bf0aa3" containerName="registry-server" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.796397 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.800603 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gr5ck"] Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.938799 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-utilities\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.939296 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5tjb\" (UniqueName: \"kubernetes.io/projected/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-kube-api-access-m5tjb\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:46 crc kubenswrapper[5035]: I1002 10:03:46.939416 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-catalog-content\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.041313 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-catalog-content\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.041471 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-utilities\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.041619 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5tjb\" (UniqueName: \"kubernetes.io/projected/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-kube-api-access-m5tjb\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.042452 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-catalog-content\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.042748 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-utilities\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.069362 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5tjb\" (UniqueName: \"kubernetes.io/projected/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-kube-api-access-m5tjb\") pod \"community-operators-gr5ck\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.134153 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:47 crc kubenswrapper[5035]: I1002 10:03:47.596734 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gr5ck"] Oct 02 10:03:47 crc kubenswrapper[5035]: W1002 10:03:47.609348 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03a71f3b_793d_4dc1_9972_3c61ac13a1d2.slice/crio-f332b19a5a1fd68090522250bcbcba8f54087cc9d6e36be0ca91b6aaccfa15fa WatchSource:0}: Error finding container f332b19a5a1fd68090522250bcbcba8f54087cc9d6e36be0ca91b6aaccfa15fa: Status 404 returned error can't find the container with id f332b19a5a1fd68090522250bcbcba8f54087cc9d6e36be0ca91b6aaccfa15fa Oct 02 10:03:48 crc kubenswrapper[5035]: I1002 10:03:48.146241 5035 generic.go:334] "Generic (PLEG): container finished" podID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerID="853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa" exitCode=0 Oct 02 10:03:48 crc kubenswrapper[5035]: I1002 10:03:48.146347 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr5ck" event={"ID":"03a71f3b-793d-4dc1-9972-3c61ac13a1d2","Type":"ContainerDied","Data":"853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa"} Oct 02 10:03:48 crc kubenswrapper[5035]: I1002 10:03:48.146593 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr5ck" event={"ID":"03a71f3b-793d-4dc1-9972-3c61ac13a1d2","Type":"ContainerStarted","Data":"f332b19a5a1fd68090522250bcbcba8f54087cc9d6e36be0ca91b6aaccfa15fa"} Oct 02 10:03:49 crc kubenswrapper[5035]: I1002 10:03:49.157860 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr5ck" event={"ID":"03a71f3b-793d-4dc1-9972-3c61ac13a1d2","Type":"ContainerStarted","Data":"6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455"} Oct 02 10:03:50 crc kubenswrapper[5035]: I1002 10:03:50.170103 5035 generic.go:334] "Generic (PLEG): container finished" podID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerID="6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455" exitCode=0 Oct 02 10:03:50 crc kubenswrapper[5035]: I1002 10:03:50.177624 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr5ck" event={"ID":"03a71f3b-793d-4dc1-9972-3c61ac13a1d2","Type":"ContainerDied","Data":"6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455"} Oct 02 10:03:51 crc kubenswrapper[5035]: I1002 10:03:51.179958 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr5ck" event={"ID":"03a71f3b-793d-4dc1-9972-3c61ac13a1d2","Type":"ContainerStarted","Data":"8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e"} Oct 02 10:03:51 crc kubenswrapper[5035]: I1002 10:03:51.210189 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gr5ck" podStartSLOduration=2.619367125 podStartE2EDuration="5.210166342s" podCreationTimestamp="2025-10-02 10:03:46 +0000 UTC" firstStartedPulling="2025-10-02 10:03:48.149459762 +0000 UTC m=+2193.505803787" lastFinishedPulling="2025-10-02 10:03:50.740258989 +0000 UTC m=+2196.096603004" observedRunningTime="2025-10-02 10:03:51.203851147 +0000 UTC m=+2196.560195172" watchObservedRunningTime="2025-10-02 10:03:51.210166342 +0000 UTC m=+2196.566510367" Oct 02 10:03:55 crc kubenswrapper[5035]: I1002 10:03:55.538438 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:03:55 crc kubenswrapper[5035]: I1002 10:03:55.539104 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:03:57 crc kubenswrapper[5035]: I1002 10:03:57.135245 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:57 crc kubenswrapper[5035]: I1002 10:03:57.135629 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:57 crc kubenswrapper[5035]: I1002 10:03:57.184884 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:57 crc kubenswrapper[5035]: I1002 10:03:57.287012 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:57 crc kubenswrapper[5035]: I1002 10:03:57.427580 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gr5ck"] Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.246287 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gr5ck" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="registry-server" containerID="cri-o://8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e" gracePeriod=2 Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.703387 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.821358 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-catalog-content\") pod \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.821496 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-utilities\") pod \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.821517 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5tjb\" (UniqueName: \"kubernetes.io/projected/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-kube-api-access-m5tjb\") pod \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\" (UID: \"03a71f3b-793d-4dc1-9972-3c61ac13a1d2\") " Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.822615 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-utilities" (OuterVolumeSpecName: "utilities") pod "03a71f3b-793d-4dc1-9972-3c61ac13a1d2" (UID: "03a71f3b-793d-4dc1-9972-3c61ac13a1d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.827761 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-kube-api-access-m5tjb" (OuterVolumeSpecName: "kube-api-access-m5tjb") pod "03a71f3b-793d-4dc1-9972-3c61ac13a1d2" (UID: "03a71f3b-793d-4dc1-9972-3c61ac13a1d2"). InnerVolumeSpecName "kube-api-access-m5tjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.874632 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03a71f3b-793d-4dc1-9972-3c61ac13a1d2" (UID: "03a71f3b-793d-4dc1-9972-3c61ac13a1d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.923456 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.923493 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5tjb\" (UniqueName: \"kubernetes.io/projected/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-kube-api-access-m5tjb\") on node \"crc\" DevicePath \"\"" Oct 02 10:03:59 crc kubenswrapper[5035]: I1002 10:03:59.923506 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03a71f3b-793d-4dc1-9972-3c61ac13a1d2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.258166 5035 generic.go:334] "Generic (PLEG): container finished" podID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerID="8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e" exitCode=0 Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.258235 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr5ck" event={"ID":"03a71f3b-793d-4dc1-9972-3c61ac13a1d2","Type":"ContainerDied","Data":"8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e"} Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.258516 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gr5ck" event={"ID":"03a71f3b-793d-4dc1-9972-3c61ac13a1d2","Type":"ContainerDied","Data":"f332b19a5a1fd68090522250bcbcba8f54087cc9d6e36be0ca91b6aaccfa15fa"} Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.258558 5035 scope.go:117] "RemoveContainer" containerID="8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.258289 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gr5ck" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.288728 5035 scope.go:117] "RemoveContainer" containerID="6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.289597 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gr5ck"] Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.296510 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gr5ck"] Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.308321 5035 scope.go:117] "RemoveContainer" containerID="853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.347436 5035 scope.go:117] "RemoveContainer" containerID="8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e" Oct 02 10:04:00 crc kubenswrapper[5035]: E1002 10:04:00.347836 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e\": container with ID starting with 8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e not found: ID does not exist" containerID="8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.347868 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e"} err="failed to get container status \"8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e\": rpc error: code = NotFound desc = could not find container \"8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e\": container with ID starting with 8e79e12e418709756b0f907bee092b8126163e204cb3ae9b54952b050bbbb72e not found: ID does not exist" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.347893 5035 scope.go:117] "RemoveContainer" containerID="6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455" Oct 02 10:04:00 crc kubenswrapper[5035]: E1002 10:04:00.348212 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455\": container with ID starting with 6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455 not found: ID does not exist" containerID="6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.348246 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455"} err="failed to get container status \"6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455\": rpc error: code = NotFound desc = could not find container \"6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455\": container with ID starting with 6116c14d139f4ccb4de1070dcba0fcf22371777d9154481a37557d1a6b8a4455 not found: ID does not exist" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.348265 5035 scope.go:117] "RemoveContainer" containerID="853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa" Oct 02 10:04:00 crc kubenswrapper[5035]: E1002 10:04:00.348556 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa\": container with ID starting with 853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa not found: ID does not exist" containerID="853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa" Oct 02 10:04:00 crc kubenswrapper[5035]: I1002 10:04:00.348584 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa"} err="failed to get container status \"853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa\": rpc error: code = NotFound desc = could not find container \"853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa\": container with ID starting with 853ef12ef624d091530494f707207f7163d553ff9cc3c8adda28964307a74baa not found: ID does not exist" Oct 02 10:04:02 crc kubenswrapper[5035]: I1002 10:04:02.183256 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" path="/var/lib/kubelet/pods/03a71f3b-793d-4dc1-9972-3c61ac13a1d2/volumes" Oct 02 10:04:25 crc kubenswrapper[5035]: I1002 10:04:25.537686 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:04:25 crc kubenswrapper[5035]: I1002 10:04:25.538708 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.537648 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.538140 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.538184 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.538960 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.539011 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" gracePeriod=600 Oct 02 10:04:55 crc kubenswrapper[5035]: E1002 10:04:55.664414 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.755040 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" exitCode=0 Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.755091 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9"} Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.755126 5035 scope.go:117] "RemoveContainer" containerID="b79bebf9853f6791ba16c2df809895df6aa63570391981a1ed44263bb7d7dbe5" Oct 02 10:04:55 crc kubenswrapper[5035]: I1002 10:04:55.755660 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:04:55 crc kubenswrapper[5035]: E1002 10:04:55.755962 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:05:11 crc kubenswrapper[5035]: I1002 10:05:11.163006 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:05:11 crc kubenswrapper[5035]: E1002 10:05:11.163872 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:05:24 crc kubenswrapper[5035]: I1002 10:05:24.164269 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:05:24 crc kubenswrapper[5035]: E1002 10:05:24.165366 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:05:39 crc kubenswrapper[5035]: I1002 10:05:39.164181 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:05:39 crc kubenswrapper[5035]: E1002 10:05:39.165382 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:05:53 crc kubenswrapper[5035]: I1002 10:05:53.163216 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:05:53 crc kubenswrapper[5035]: E1002 10:05:53.164003 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:06:04 crc kubenswrapper[5035]: I1002 10:06:04.164095 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:06:04 crc kubenswrapper[5035]: E1002 10:06:04.164985 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:06:19 crc kubenswrapper[5035]: I1002 10:06:19.163862 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:06:19 crc kubenswrapper[5035]: E1002 10:06:19.164605 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:06:34 crc kubenswrapper[5035]: I1002 10:06:34.163685 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:06:34 crc kubenswrapper[5035]: E1002 10:06:34.164696 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:06:46 crc kubenswrapper[5035]: I1002 10:06:46.182235 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:06:46 crc kubenswrapper[5035]: E1002 10:06:46.184023 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:06:51 crc kubenswrapper[5035]: I1002 10:06:51.919147 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7cnwl"] Oct 02 10:06:51 crc kubenswrapper[5035]: E1002 10:06:51.920097 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="extract-content" Oct 02 10:06:51 crc kubenswrapper[5035]: I1002 10:06:51.920109 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="extract-content" Oct 02 10:06:51 crc kubenswrapper[5035]: E1002 10:06:51.920128 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="extract-utilities" Oct 02 10:06:51 crc kubenswrapper[5035]: I1002 10:06:51.920135 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="extract-utilities" Oct 02 10:06:51 crc kubenswrapper[5035]: E1002 10:06:51.920146 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="registry-server" Oct 02 10:06:51 crc kubenswrapper[5035]: I1002 10:06:51.920153 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="registry-server" Oct 02 10:06:51 crc kubenswrapper[5035]: I1002 10:06:51.920363 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="03a71f3b-793d-4dc1-9972-3c61ac13a1d2" containerName="registry-server" Oct 02 10:06:51 crc kubenswrapper[5035]: I1002 10:06:51.921829 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:51 crc kubenswrapper[5035]: I1002 10:06:51.941971 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7cnwl"] Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.047210 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ctdb\" (UniqueName: \"kubernetes.io/projected/874c5bd8-63f0-4781-8709-c07fac93d8da-kube-api-access-9ctdb\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.047331 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-utilities\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.047626 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-catalog-content\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.149685 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ctdb\" (UniqueName: \"kubernetes.io/projected/874c5bd8-63f0-4781-8709-c07fac93d8da-kube-api-access-9ctdb\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.150009 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-utilities\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.150062 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-catalog-content\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.150668 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-catalog-content\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.150752 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-utilities\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.172796 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ctdb\" (UniqueName: \"kubernetes.io/projected/874c5bd8-63f0-4781-8709-c07fac93d8da-kube-api-access-9ctdb\") pod \"certified-operators-7cnwl\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.246052 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.751263 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7cnwl"] Oct 02 10:06:52 crc kubenswrapper[5035]: I1002 10:06:52.798391 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cnwl" event={"ID":"874c5bd8-63f0-4781-8709-c07fac93d8da","Type":"ContainerStarted","Data":"3276eb355823a68c22777a59d9071b09f67f2b36f6e2143c6d8919cdb3d7d9bc"} Oct 02 10:06:53 crc kubenswrapper[5035]: I1002 10:06:53.809285 5035 generic.go:334] "Generic (PLEG): container finished" podID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerID="227c18356547415293afd389642678968133249bd7973188575a3339ebc07410" exitCode=0 Oct 02 10:06:53 crc kubenswrapper[5035]: I1002 10:06:53.809405 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cnwl" event={"ID":"874c5bd8-63f0-4781-8709-c07fac93d8da","Type":"ContainerDied","Data":"227c18356547415293afd389642678968133249bd7973188575a3339ebc07410"} Oct 02 10:06:55 crc kubenswrapper[5035]: I1002 10:06:55.841592 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cnwl" event={"ID":"874c5bd8-63f0-4781-8709-c07fac93d8da","Type":"ContainerDied","Data":"10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9"} Oct 02 10:06:55 crc kubenswrapper[5035]: I1002 10:06:55.843679 5035 generic.go:334] "Generic (PLEG): container finished" podID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerID="10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9" exitCode=0 Oct 02 10:06:56 crc kubenswrapper[5035]: I1002 10:06:56.864346 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cnwl" event={"ID":"874c5bd8-63f0-4781-8709-c07fac93d8da","Type":"ContainerStarted","Data":"4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5"} Oct 02 10:06:59 crc kubenswrapper[5035]: I1002 10:06:59.163249 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:06:59 crc kubenswrapper[5035]: E1002 10:06:59.163832 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:07:02 crc kubenswrapper[5035]: I1002 10:07:02.246459 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:07:02 crc kubenswrapper[5035]: I1002 10:07:02.246807 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:07:02 crc kubenswrapper[5035]: I1002 10:07:02.292450 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:07:02 crc kubenswrapper[5035]: I1002 10:07:02.313153 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7cnwl" podStartSLOduration=8.886465041 podStartE2EDuration="11.313131409s" podCreationTimestamp="2025-10-02 10:06:51 +0000 UTC" firstStartedPulling="2025-10-02 10:06:53.811443776 +0000 UTC m=+2379.167787801" lastFinishedPulling="2025-10-02 10:06:56.238110144 +0000 UTC m=+2381.594454169" observedRunningTime="2025-10-02 10:06:56.882739946 +0000 UTC m=+2382.239084011" watchObservedRunningTime="2025-10-02 10:07:02.313131409 +0000 UTC m=+2387.669475444" Oct 02 10:07:02 crc kubenswrapper[5035]: I1002 10:07:02.972697 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:07:03 crc kubenswrapper[5035]: I1002 10:07:03.019835 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7cnwl"] Oct 02 10:07:04 crc kubenswrapper[5035]: I1002 10:07:04.942505 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7cnwl" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="registry-server" containerID="cri-o://4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5" gracePeriod=2 Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.323526 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.424812 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ctdb\" (UniqueName: \"kubernetes.io/projected/874c5bd8-63f0-4781-8709-c07fac93d8da-kube-api-access-9ctdb\") pod \"874c5bd8-63f0-4781-8709-c07fac93d8da\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.424912 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-catalog-content\") pod \"874c5bd8-63f0-4781-8709-c07fac93d8da\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.425288 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-utilities\") pod \"874c5bd8-63f0-4781-8709-c07fac93d8da\" (UID: \"874c5bd8-63f0-4781-8709-c07fac93d8da\") " Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.426445 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-utilities" (OuterVolumeSpecName: "utilities") pod "874c5bd8-63f0-4781-8709-c07fac93d8da" (UID: "874c5bd8-63f0-4781-8709-c07fac93d8da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.432632 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/874c5bd8-63f0-4781-8709-c07fac93d8da-kube-api-access-9ctdb" (OuterVolumeSpecName: "kube-api-access-9ctdb") pod "874c5bd8-63f0-4781-8709-c07fac93d8da" (UID: "874c5bd8-63f0-4781-8709-c07fac93d8da"). InnerVolumeSpecName "kube-api-access-9ctdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.528041 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.528077 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ctdb\" (UniqueName: \"kubernetes.io/projected/874c5bd8-63f0-4781-8709-c07fac93d8da-kube-api-access-9ctdb\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.556794 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "874c5bd8-63f0-4781-8709-c07fac93d8da" (UID: "874c5bd8-63f0-4781-8709-c07fac93d8da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.629691 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/874c5bd8-63f0-4781-8709-c07fac93d8da-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.954627 5035 generic.go:334] "Generic (PLEG): container finished" podID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerID="4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5" exitCode=0 Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.954676 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cnwl" event={"ID":"874c5bd8-63f0-4781-8709-c07fac93d8da","Type":"ContainerDied","Data":"4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5"} Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.954705 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7cnwl" event={"ID":"874c5bd8-63f0-4781-8709-c07fac93d8da","Type":"ContainerDied","Data":"3276eb355823a68c22777a59d9071b09f67f2b36f6e2143c6d8919cdb3d7d9bc"} Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.954713 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7cnwl" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.954724 5035 scope.go:117] "RemoveContainer" containerID="4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.976788 5035 scope.go:117] "RemoveContainer" containerID="10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9" Oct 02 10:07:05 crc kubenswrapper[5035]: I1002 10:07:05.994724 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7cnwl"] Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.004219 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7cnwl"] Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.029294 5035 scope.go:117] "RemoveContainer" containerID="227c18356547415293afd389642678968133249bd7973188575a3339ebc07410" Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.061245 5035 scope.go:117] "RemoveContainer" containerID="4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5" Oct 02 10:07:06 crc kubenswrapper[5035]: E1002 10:07:06.061827 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5\": container with ID starting with 4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5 not found: ID does not exist" containerID="4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5" Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.061883 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5"} err="failed to get container status \"4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5\": rpc error: code = NotFound desc = could not find container \"4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5\": container with ID starting with 4fd115429a534c5b042411dae98edff5b16130b5f54665c12307df0c48d9bad5 not found: ID does not exist" Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.061910 5035 scope.go:117] "RemoveContainer" containerID="10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9" Oct 02 10:07:06 crc kubenswrapper[5035]: E1002 10:07:06.063022 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9\": container with ID starting with 10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9 not found: ID does not exist" containerID="10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9" Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.063049 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9"} err="failed to get container status \"10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9\": rpc error: code = NotFound desc = could not find container \"10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9\": container with ID starting with 10438341749f28cde25bb506516670606914277be00d726d01655d87baf462e9 not found: ID does not exist" Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.063065 5035 scope.go:117] "RemoveContainer" containerID="227c18356547415293afd389642678968133249bd7973188575a3339ebc07410" Oct 02 10:07:06 crc kubenswrapper[5035]: E1002 10:07:06.063340 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"227c18356547415293afd389642678968133249bd7973188575a3339ebc07410\": container with ID starting with 227c18356547415293afd389642678968133249bd7973188575a3339ebc07410 not found: ID does not exist" containerID="227c18356547415293afd389642678968133249bd7973188575a3339ebc07410" Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.063384 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"227c18356547415293afd389642678968133249bd7973188575a3339ebc07410"} err="failed to get container status \"227c18356547415293afd389642678968133249bd7973188575a3339ebc07410\": rpc error: code = NotFound desc = could not find container \"227c18356547415293afd389642678968133249bd7973188575a3339ebc07410\": container with ID starting with 227c18356547415293afd389642678968133249bd7973188575a3339ebc07410 not found: ID does not exist" Oct 02 10:07:06 crc kubenswrapper[5035]: I1002 10:07:06.180447 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" path="/var/lib/kubelet/pods/874c5bd8-63f0-4781-8709-c07fac93d8da/volumes" Oct 02 10:07:13 crc kubenswrapper[5035]: I1002 10:07:13.163974 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:07:13 crc kubenswrapper[5035]: E1002 10:07:13.165157 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:07:25 crc kubenswrapper[5035]: I1002 10:07:25.163329 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:07:25 crc kubenswrapper[5035]: E1002 10:07:25.164494 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:07:39 crc kubenswrapper[5035]: I1002 10:07:39.163197 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:07:39 crc kubenswrapper[5035]: E1002 10:07:39.163973 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:07:52 crc kubenswrapper[5035]: I1002 10:07:52.163782 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:07:52 crc kubenswrapper[5035]: E1002 10:07:52.164843 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:08:04 crc kubenswrapper[5035]: I1002 10:08:04.163099 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:08:04 crc kubenswrapper[5035]: E1002 10:08:04.163850 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:08:18 crc kubenswrapper[5035]: I1002 10:08:18.163406 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:08:18 crc kubenswrapper[5035]: E1002 10:08:18.164257 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:08:30 crc kubenswrapper[5035]: I1002 10:08:30.162987 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:08:30 crc kubenswrapper[5035]: E1002 10:08:30.164708 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:08:41 crc kubenswrapper[5035]: I1002 10:08:41.162749 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:08:41 crc kubenswrapper[5035]: E1002 10:08:41.163494 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:08:53 crc kubenswrapper[5035]: I1002 10:08:53.162770 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:08:53 crc kubenswrapper[5035]: E1002 10:08:53.163570 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:09:06 crc kubenswrapper[5035]: I1002 10:09:06.171568 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:09:06 crc kubenswrapper[5035]: E1002 10:09:06.172579 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:09:20 crc kubenswrapper[5035]: I1002 10:09:20.162581 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:09:20 crc kubenswrapper[5035]: E1002 10:09:20.163398 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:09:34 crc kubenswrapper[5035]: I1002 10:09:34.162850 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:09:34 crc kubenswrapper[5035]: E1002 10:09:34.163653 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:09:46 crc kubenswrapper[5035]: I1002 10:09:46.170185 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:09:46 crc kubenswrapper[5035]: E1002 10:09:46.170944 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:09:59 crc kubenswrapper[5035]: I1002 10:09:59.163057 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:09:59 crc kubenswrapper[5035]: I1002 10:09:59.532772 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"668bb58c600ac20dc33ddeae8555bcaa9ee1a27b46e757cd7dc4e3dbed851ad5"} Oct 02 10:12:25 crc kubenswrapper[5035]: I1002 10:12:25.539384 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:12:25 crc kubenswrapper[5035]: I1002 10:12:25.540206 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:12:55 crc kubenswrapper[5035]: I1002 10:12:55.538401 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:12:55 crc kubenswrapper[5035]: I1002 10:12:55.539269 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:13:25 crc kubenswrapper[5035]: I1002 10:13:25.538128 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:13:25 crc kubenswrapper[5035]: I1002 10:13:25.538746 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:13:25 crc kubenswrapper[5035]: I1002 10:13:25.538796 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 10:13:25 crc kubenswrapper[5035]: I1002 10:13:25.539303 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"668bb58c600ac20dc33ddeae8555bcaa9ee1a27b46e757cd7dc4e3dbed851ad5"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:13:25 crc kubenswrapper[5035]: I1002 10:13:25.539364 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://668bb58c600ac20dc33ddeae8555bcaa9ee1a27b46e757cd7dc4e3dbed851ad5" gracePeriod=600 Oct 02 10:13:26 crc kubenswrapper[5035]: I1002 10:13:26.514765 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="668bb58c600ac20dc33ddeae8555bcaa9ee1a27b46e757cd7dc4e3dbed851ad5" exitCode=0 Oct 02 10:13:26 crc kubenswrapper[5035]: I1002 10:13:26.514830 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"668bb58c600ac20dc33ddeae8555bcaa9ee1a27b46e757cd7dc4e3dbed851ad5"} Oct 02 10:13:26 crc kubenswrapper[5035]: I1002 10:13:26.515580 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerStarted","Data":"a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3"} Oct 02 10:13:26 crc kubenswrapper[5035]: I1002 10:13:26.515616 5035 scope.go:117] "RemoveContainer" containerID="320959d35ddb1634b43c72277fd4706f1392dd8a477bf758238e6cf1640331e9" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.048853 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8prlx"] Oct 02 10:14:05 crc kubenswrapper[5035]: E1002 10:14:05.050496 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="extract-utilities" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.050517 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="extract-utilities" Oct 02 10:14:05 crc kubenswrapper[5035]: E1002 10:14:05.050557 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="extract-content" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.050564 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="extract-content" Oct 02 10:14:05 crc kubenswrapper[5035]: E1002 10:14:05.050577 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="registry-server" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.050585 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="registry-server" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.050939 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="874c5bd8-63f0-4781-8709-c07fac93d8da" containerName="registry-server" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.052926 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.106368 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prlx"] Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.156869 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-utilities\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.156967 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-catalog-content\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.157080 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4c45\" (UniqueName: \"kubernetes.io/projected/5b76b5c8-bf44-4dec-8b57-02a21074a44b-kube-api-access-l4c45\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.260016 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-utilities\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.260202 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-catalog-content\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.260586 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-utilities\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.260726 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-catalog-content\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.260754 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4c45\" (UniqueName: \"kubernetes.io/projected/5b76b5c8-bf44-4dec-8b57-02a21074a44b-kube-api-access-l4c45\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.293022 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4c45\" (UniqueName: \"kubernetes.io/projected/5b76b5c8-bf44-4dec-8b57-02a21074a44b-kube-api-access-l4c45\") pod \"redhat-marketplace-8prlx\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.430018 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:05 crc kubenswrapper[5035]: I1002 10:14:05.919763 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prlx"] Oct 02 10:14:06 crc kubenswrapper[5035]: I1002 10:14:06.960770 5035 generic.go:334] "Generic (PLEG): container finished" podID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerID="fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1" exitCode=0 Oct 02 10:14:06 crc kubenswrapper[5035]: I1002 10:14:06.961251 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prlx" event={"ID":"5b76b5c8-bf44-4dec-8b57-02a21074a44b","Type":"ContainerDied","Data":"fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1"} Oct 02 10:14:06 crc kubenswrapper[5035]: I1002 10:14:06.961300 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prlx" event={"ID":"5b76b5c8-bf44-4dec-8b57-02a21074a44b","Type":"ContainerStarted","Data":"cad2ff4d6ec9c5adaec7c283e933a8176f8209be381a887bca954781bfa34c9d"} Oct 02 10:14:06 crc kubenswrapper[5035]: I1002 10:14:06.964797 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:14:07 crc kubenswrapper[5035]: I1002 10:14:07.971181 5035 generic.go:334] "Generic (PLEG): container finished" podID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerID="3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7" exitCode=0 Oct 02 10:14:07 crc kubenswrapper[5035]: I1002 10:14:07.971321 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prlx" event={"ID":"5b76b5c8-bf44-4dec-8b57-02a21074a44b","Type":"ContainerDied","Data":"3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7"} Oct 02 10:14:08 crc kubenswrapper[5035]: I1002 10:14:08.984987 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prlx" event={"ID":"5b76b5c8-bf44-4dec-8b57-02a21074a44b","Type":"ContainerStarted","Data":"86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea"} Oct 02 10:14:09 crc kubenswrapper[5035]: I1002 10:14:09.007578 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8prlx" podStartSLOduration=2.278309619 podStartE2EDuration="4.007525363s" podCreationTimestamp="2025-10-02 10:14:05 +0000 UTC" firstStartedPulling="2025-10-02 10:14:06.964353834 +0000 UTC m=+2812.320697879" lastFinishedPulling="2025-10-02 10:14:08.693569608 +0000 UTC m=+2814.049913623" observedRunningTime="2025-10-02 10:14:09.006363259 +0000 UTC m=+2814.362707284" watchObservedRunningTime="2025-10-02 10:14:09.007525363 +0000 UTC m=+2814.363869398" Oct 02 10:14:15 crc kubenswrapper[5035]: I1002 10:14:15.431160 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:15 crc kubenswrapper[5035]: I1002 10:14:15.431788 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:15 crc kubenswrapper[5035]: I1002 10:14:15.489699 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:16 crc kubenswrapper[5035]: I1002 10:14:16.121672 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:16 crc kubenswrapper[5035]: I1002 10:14:16.182582 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prlx"] Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.086607 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8prlx" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="registry-server" containerID="cri-o://86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea" gracePeriod=2 Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.589957 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.651755 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4c45\" (UniqueName: \"kubernetes.io/projected/5b76b5c8-bf44-4dec-8b57-02a21074a44b-kube-api-access-l4c45\") pod \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.651804 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-utilities\") pod \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.653128 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-utilities" (OuterVolumeSpecName: "utilities") pod "5b76b5c8-bf44-4dec-8b57-02a21074a44b" (UID: "5b76b5c8-bf44-4dec-8b57-02a21074a44b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.653194 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-catalog-content\") pod \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\" (UID: \"5b76b5c8-bf44-4dec-8b57-02a21074a44b\") " Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.654283 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.658234 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b76b5c8-bf44-4dec-8b57-02a21074a44b-kube-api-access-l4c45" (OuterVolumeSpecName: "kube-api-access-l4c45") pod "5b76b5c8-bf44-4dec-8b57-02a21074a44b" (UID: "5b76b5c8-bf44-4dec-8b57-02a21074a44b"). InnerVolumeSpecName "kube-api-access-l4c45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.671212 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b76b5c8-bf44-4dec-8b57-02a21074a44b" (UID: "5b76b5c8-bf44-4dec-8b57-02a21074a44b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.756324 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4c45\" (UniqueName: \"kubernetes.io/projected/5b76b5c8-bf44-4dec-8b57-02a21074a44b-kube-api-access-l4c45\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:18 crc kubenswrapper[5035]: I1002 10:14:18.756374 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b76b5c8-bf44-4dec-8b57-02a21074a44b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.098185 5035 generic.go:334] "Generic (PLEG): container finished" podID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerID="86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea" exitCode=0 Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.098254 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prlx" event={"ID":"5b76b5c8-bf44-4dec-8b57-02a21074a44b","Type":"ContainerDied","Data":"86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea"} Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.098289 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8prlx" event={"ID":"5b76b5c8-bf44-4dec-8b57-02a21074a44b","Type":"ContainerDied","Data":"cad2ff4d6ec9c5adaec7c283e933a8176f8209be381a887bca954781bfa34c9d"} Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.098292 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8prlx" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.098311 5035 scope.go:117] "RemoveContainer" containerID="86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.126495 5035 scope.go:117] "RemoveContainer" containerID="3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.160355 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prlx"] Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.168716 5035 scope.go:117] "RemoveContainer" containerID="fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.174851 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8prlx"] Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.213725 5035 scope.go:117] "RemoveContainer" containerID="86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea" Oct 02 10:14:19 crc kubenswrapper[5035]: E1002 10:14:19.214325 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea\": container with ID starting with 86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea not found: ID does not exist" containerID="86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.214358 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea"} err="failed to get container status \"86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea\": rpc error: code = NotFound desc = could not find container \"86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea\": container with ID starting with 86e727e99aec41e233ae99dc2cb67a78c94d74f15df5c2eb0a2fca49d6128aea not found: ID does not exist" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.214382 5035 scope.go:117] "RemoveContainer" containerID="3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7" Oct 02 10:14:19 crc kubenswrapper[5035]: E1002 10:14:19.214760 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7\": container with ID starting with 3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7 not found: ID does not exist" containerID="3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.214830 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7"} err="failed to get container status \"3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7\": rpc error: code = NotFound desc = could not find container \"3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7\": container with ID starting with 3158552972b2563ee2f607358abf4b6395fcafe7fd9ebfd903f5ffa0f5874fa7 not found: ID does not exist" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.214877 5035 scope.go:117] "RemoveContainer" containerID="fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1" Oct 02 10:14:19 crc kubenswrapper[5035]: E1002 10:14:19.215308 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1\": container with ID starting with fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1 not found: ID does not exist" containerID="fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.215370 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1"} err="failed to get container status \"fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1\": rpc error: code = NotFound desc = could not find container \"fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1\": container with ID starting with fea64b79a8f7dd8c9ecda64066695bd51331a3fde6f1c497d7ed3ea5fe1f41a1 not found: ID does not exist" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.709349 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q978g/must-gather-8ncgt"] Oct 02 10:14:19 crc kubenswrapper[5035]: E1002 10:14:19.710143 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="extract-utilities" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.710173 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="extract-utilities" Oct 02 10:14:19 crc kubenswrapper[5035]: E1002 10:14:19.710198 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="registry-server" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.710207 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="registry-server" Oct 02 10:14:19 crc kubenswrapper[5035]: E1002 10:14:19.710232 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="extract-content" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.710241 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="extract-content" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.710569 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" containerName="registry-server" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.711854 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.716026 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-q978g"/"default-dockercfg-42z8v" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.716320 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-q978g"/"kube-root-ca.crt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.716584 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-q978g"/"openshift-service-ca.crt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.738610 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-q978g/must-gather-8ncgt"] Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.784904 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ng6s6\" (UniqueName: \"kubernetes.io/projected/183893c4-7c7f-4e8a-80a0-5850149d5f9c-kube-api-access-ng6s6\") pod \"must-gather-8ncgt\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.784975 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/183893c4-7c7f-4e8a-80a0-5850149d5f9c-must-gather-output\") pod \"must-gather-8ncgt\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.886194 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ng6s6\" (UniqueName: \"kubernetes.io/projected/183893c4-7c7f-4e8a-80a0-5850149d5f9c-kube-api-access-ng6s6\") pod \"must-gather-8ncgt\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.886276 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/183893c4-7c7f-4e8a-80a0-5850149d5f9c-must-gather-output\") pod \"must-gather-8ncgt\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.886944 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/183893c4-7c7f-4e8a-80a0-5850149d5f9c-must-gather-output\") pod \"must-gather-8ncgt\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:19 crc kubenswrapper[5035]: I1002 10:14:19.912831 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ng6s6\" (UniqueName: \"kubernetes.io/projected/183893c4-7c7f-4e8a-80a0-5850149d5f9c-kube-api-access-ng6s6\") pod \"must-gather-8ncgt\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:20 crc kubenswrapper[5035]: I1002 10:14:20.041833 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:14:20 crc kubenswrapper[5035]: I1002 10:14:20.176224 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b76b5c8-bf44-4dec-8b57-02a21074a44b" path="/var/lib/kubelet/pods/5b76b5c8-bf44-4dec-8b57-02a21074a44b/volumes" Oct 02 10:14:20 crc kubenswrapper[5035]: I1002 10:14:20.642301 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-q978g/must-gather-8ncgt"] Oct 02 10:14:20 crc kubenswrapper[5035]: W1002 10:14:20.644909 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod183893c4_7c7f_4e8a_80a0_5850149d5f9c.slice/crio-3903ddf1b42474eed0d6ae3503d779997f525a28a849f6b93c200d0edba4baa7 WatchSource:0}: Error finding container 3903ddf1b42474eed0d6ae3503d779997f525a28a849f6b93c200d0edba4baa7: Status 404 returned error can't find the container with id 3903ddf1b42474eed0d6ae3503d779997f525a28a849f6b93c200d0edba4baa7 Oct 02 10:14:21 crc kubenswrapper[5035]: I1002 10:14:21.126270 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/must-gather-8ncgt" event={"ID":"183893c4-7c7f-4e8a-80a0-5850149d5f9c","Type":"ContainerStarted","Data":"3903ddf1b42474eed0d6ae3503d779997f525a28a849f6b93c200d0edba4baa7"} Oct 02 10:14:28 crc kubenswrapper[5035]: I1002 10:14:28.215954 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/must-gather-8ncgt" event={"ID":"183893c4-7c7f-4e8a-80a0-5850149d5f9c","Type":"ContainerStarted","Data":"0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2"} Oct 02 10:14:28 crc kubenswrapper[5035]: I1002 10:14:28.216359 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/must-gather-8ncgt" event={"ID":"183893c4-7c7f-4e8a-80a0-5850149d5f9c","Type":"ContainerStarted","Data":"89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b"} Oct 02 10:14:28 crc kubenswrapper[5035]: I1002 10:14:28.235574 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-q978g/must-gather-8ncgt" podStartSLOduration=2.719411138 podStartE2EDuration="9.235556731s" podCreationTimestamp="2025-10-02 10:14:19 +0000 UTC" firstStartedPulling="2025-10-02 10:14:20.648177423 +0000 UTC m=+2826.004521448" lastFinishedPulling="2025-10-02 10:14:27.164323016 +0000 UTC m=+2832.520667041" observedRunningTime="2025-10-02 10:14:28.229825693 +0000 UTC m=+2833.586169718" watchObservedRunningTime="2025-10-02 10:14:28.235556731 +0000 UTC m=+2833.591900756" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.417156 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q978g/crc-debug-xj978"] Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.427285 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.565599 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fkx9\" (UniqueName: \"kubernetes.io/projected/1e40b571-0063-4584-b95b-f756e1976f89-kube-api-access-2fkx9\") pod \"crc-debug-xj978\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.566178 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e40b571-0063-4584-b95b-f756e1976f89-host\") pod \"crc-debug-xj978\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.668750 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fkx9\" (UniqueName: \"kubernetes.io/projected/1e40b571-0063-4584-b95b-f756e1976f89-kube-api-access-2fkx9\") pod \"crc-debug-xj978\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.668929 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e40b571-0063-4584-b95b-f756e1976f89-host\") pod \"crc-debug-xj978\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.669125 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e40b571-0063-4584-b95b-f756e1976f89-host\") pod \"crc-debug-xj978\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.688040 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fkx9\" (UniqueName: \"kubernetes.io/projected/1e40b571-0063-4584-b95b-f756e1976f89-kube-api-access-2fkx9\") pod \"crc-debug-xj978\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:32 crc kubenswrapper[5035]: I1002 10:14:32.759781 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:14:33 crc kubenswrapper[5035]: I1002 10:14:33.259737 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-xj978" event={"ID":"1e40b571-0063-4584-b95b-f756e1976f89","Type":"ContainerStarted","Data":"4ce4c3716238f95839d9c247927d4279b00ef3f3dde5c8c04fb3807b57c69d76"} Oct 02 10:14:44 crc kubenswrapper[5035]: I1002 10:14:44.359021 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-xj978" event={"ID":"1e40b571-0063-4584-b95b-f756e1976f89","Type":"ContainerStarted","Data":"87e5d50e2ac405750b45e0d3f869bfd3e0d7745faf11d8b8de51be31eaa41e4f"} Oct 02 10:14:44 crc kubenswrapper[5035]: I1002 10:14:44.381373 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-q978g/crc-debug-xj978" podStartSLOduration=1.230378537 podStartE2EDuration="12.381349925s" podCreationTimestamp="2025-10-02 10:14:32 +0000 UTC" firstStartedPulling="2025-10-02 10:14:32.795943355 +0000 UTC m=+2838.152287390" lastFinishedPulling="2025-10-02 10:14:43.946914753 +0000 UTC m=+2849.303258778" observedRunningTime="2025-10-02 10:14:44.375781222 +0000 UTC m=+2849.732125267" watchObservedRunningTime="2025-10-02 10:14:44.381349925 +0000 UTC m=+2849.737693950" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.141741 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw"] Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.143479 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.146472 5035 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.146572 5035 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.180860 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw"] Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.220898 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p89rq\" (UniqueName: \"kubernetes.io/projected/7d060000-b611-4459-b4f4-90dbf4835f12-kube-api-access-p89rq\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.221353 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d060000-b611-4459-b4f4-90dbf4835f12-secret-volume\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.221462 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d060000-b611-4459-b4f4-90dbf4835f12-config-volume\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.323243 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d060000-b611-4459-b4f4-90dbf4835f12-secret-volume\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.323339 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d060000-b611-4459-b4f4-90dbf4835f12-config-volume\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.323449 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p89rq\" (UniqueName: \"kubernetes.io/projected/7d060000-b611-4459-b4f4-90dbf4835f12-kube-api-access-p89rq\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.324703 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d060000-b611-4459-b4f4-90dbf4835f12-config-volume\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.330859 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d060000-b611-4459-b4f4-90dbf4835f12-secret-volume\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.346135 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p89rq\" (UniqueName: \"kubernetes.io/projected/7d060000-b611-4459-b4f4-90dbf4835f12-kube-api-access-p89rq\") pod \"collect-profiles-29323335-knhrw\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.467604 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:00 crc kubenswrapper[5035]: I1002 10:15:00.959999 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw"] Oct 02 10:15:00 crc kubenswrapper[5035]: W1002 10:15:00.977250 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d060000_b611_4459_b4f4_90dbf4835f12.slice/crio-1c891648be89b72dac9eba70874b3d01bbfef394a4a36c7ce062ea75758e9e75 WatchSource:0}: Error finding container 1c891648be89b72dac9eba70874b3d01bbfef394a4a36c7ce062ea75758e9e75: Status 404 returned error can't find the container with id 1c891648be89b72dac9eba70874b3d01bbfef394a4a36c7ce062ea75758e9e75 Oct 02 10:15:01 crc kubenswrapper[5035]: I1002 10:15:01.531202 5035 generic.go:334] "Generic (PLEG): container finished" podID="7d060000-b611-4459-b4f4-90dbf4835f12" containerID="34f820a0af4ca1e937d262594f98b706bbf33a80111d76862c899a58aefd07a9" exitCode=0 Oct 02 10:15:01 crc kubenswrapper[5035]: I1002 10:15:01.531429 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" event={"ID":"7d060000-b611-4459-b4f4-90dbf4835f12","Type":"ContainerDied","Data":"34f820a0af4ca1e937d262594f98b706bbf33a80111d76862c899a58aefd07a9"} Oct 02 10:15:01 crc kubenswrapper[5035]: I1002 10:15:01.531453 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" event={"ID":"7d060000-b611-4459-b4f4-90dbf4835f12","Type":"ContainerStarted","Data":"1c891648be89b72dac9eba70874b3d01bbfef394a4a36c7ce062ea75758e9e75"} Oct 02 10:15:02 crc kubenswrapper[5035]: I1002 10:15:02.885450 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:02 crc kubenswrapper[5035]: I1002 10:15:02.980689 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d060000-b611-4459-b4f4-90dbf4835f12-secret-volume\") pod \"7d060000-b611-4459-b4f4-90dbf4835f12\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " Oct 02 10:15:02 crc kubenswrapper[5035]: I1002 10:15:02.980768 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p89rq\" (UniqueName: \"kubernetes.io/projected/7d060000-b611-4459-b4f4-90dbf4835f12-kube-api-access-p89rq\") pod \"7d060000-b611-4459-b4f4-90dbf4835f12\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " Oct 02 10:15:02 crc kubenswrapper[5035]: I1002 10:15:02.980805 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d060000-b611-4459-b4f4-90dbf4835f12-config-volume\") pod \"7d060000-b611-4459-b4f4-90dbf4835f12\" (UID: \"7d060000-b611-4459-b4f4-90dbf4835f12\") " Oct 02 10:15:02 crc kubenswrapper[5035]: I1002 10:15:02.982118 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d060000-b611-4459-b4f4-90dbf4835f12-config-volume" (OuterVolumeSpecName: "config-volume") pod "7d060000-b611-4459-b4f4-90dbf4835f12" (UID: "7d060000-b611-4459-b4f4-90dbf4835f12"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 10:15:02 crc kubenswrapper[5035]: I1002 10:15:02.986473 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d060000-b611-4459-b4f4-90dbf4835f12-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7d060000-b611-4459-b4f4-90dbf4835f12" (UID: "7d060000-b611-4459-b4f4-90dbf4835f12"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 10:15:02 crc kubenswrapper[5035]: I1002 10:15:02.986724 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d060000-b611-4459-b4f4-90dbf4835f12-kube-api-access-p89rq" (OuterVolumeSpecName: "kube-api-access-p89rq") pod "7d060000-b611-4459-b4f4-90dbf4835f12" (UID: "7d060000-b611-4459-b4f4-90dbf4835f12"). InnerVolumeSpecName "kube-api-access-p89rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.083205 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p89rq\" (UniqueName: \"kubernetes.io/projected/7d060000-b611-4459-b4f4-90dbf4835f12-kube-api-access-p89rq\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.083251 5035 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d060000-b611-4459-b4f4-90dbf4835f12-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.083264 5035 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d060000-b611-4459-b4f4-90dbf4835f12-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.553673 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" event={"ID":"7d060000-b611-4459-b4f4-90dbf4835f12","Type":"ContainerDied","Data":"1c891648be89b72dac9eba70874b3d01bbfef394a4a36c7ce062ea75758e9e75"} Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.553718 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c891648be89b72dac9eba70874b3d01bbfef394a4a36c7ce062ea75758e9e75" Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.553829 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323335-knhrw" Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.958707 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465"] Oct 02 10:15:03 crc kubenswrapper[5035]: I1002 10:15:03.966025 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323290-7w465"] Oct 02 10:15:04 crc kubenswrapper[5035]: I1002 10:15:04.174392 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9af7278-65f4-471e-8b74-4305e9174038" path="/var/lib/kubelet/pods/e9af7278-65f4-471e-8b74-4305e9174038/volumes" Oct 02 10:15:25 crc kubenswrapper[5035]: I1002 10:15:25.537850 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:15:25 crc kubenswrapper[5035]: I1002 10:15:25.538557 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:15:28 crc kubenswrapper[5035]: I1002 10:15:28.711348 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7975d57bf8-2hqjm_78fda780-4464-412b-919b-58ee75780eb6/barbican-api-log/0.log" Oct 02 10:15:28 crc kubenswrapper[5035]: I1002 10:15:28.719849 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7975d57bf8-2hqjm_78fda780-4464-412b-919b-58ee75780eb6/barbican-api/0.log" Oct 02 10:15:28 crc kubenswrapper[5035]: I1002 10:15:28.892644 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-545df9b758-drjst_54742b9b-0346-4cdb-9bf2-75f35c59beaf/barbican-keystone-listener/0.log" Oct 02 10:15:28 crc kubenswrapper[5035]: I1002 10:15:28.966738 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-545df9b758-drjst_54742b9b-0346-4cdb-9bf2-75f35c59beaf/barbican-keystone-listener-log/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.191672 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d9b9799bc-tvwd9_76945d28-7017-4b37-ac66-c356e9190230/barbican-worker-log/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.227871 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5d9b9799bc-tvwd9_76945d28-7017-4b37-ac66-c356e9190230/barbican-worker/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.473606 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0400439-a652-4472-9f8a-51323ea10780/ceilometer-central-agent/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.503093 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0400439-a652-4472-9f8a-51323ea10780/ceilometer-notification-agent/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.683685 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0400439-a652-4472-9f8a-51323ea10780/proxy-httpd/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.745836 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_a0400439-a652-4472-9f8a-51323ea10780/sg-core/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.940454 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8ccd0633-b703-46d1-b74c-9129d3b5b9ac/cinder-api/0.log" Oct 02 10:15:29 crc kubenswrapper[5035]: I1002 10:15:29.983071 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8ccd0633-b703-46d1-b74c-9129d3b5b9ac/cinder-api-log/0.log" Oct 02 10:15:30 crc kubenswrapper[5035]: I1002 10:15:30.200315 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_624fbc6b-e52c-4d61-932e-24c8a5686a25/probe/0.log" Oct 02 10:15:30 crc kubenswrapper[5035]: I1002 10:15:30.204560 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_624fbc6b-e52c-4d61-932e-24c8a5686a25/cinder-scheduler/0.log" Oct 02 10:15:30 crc kubenswrapper[5035]: I1002 10:15:30.390252 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-59cf4bdb65-lmk4b_85b5680e-bfe0-4c34-b773-14364d25cac5/init/0.log" Oct 02 10:15:30 crc kubenswrapper[5035]: I1002 10:15:30.596896 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-59cf4bdb65-lmk4b_85b5680e-bfe0-4c34-b773-14364d25cac5/dnsmasq-dns/0.log" Oct 02 10:15:30 crc kubenswrapper[5035]: I1002 10:15:30.619406 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-59cf4bdb65-lmk4b_85b5680e-bfe0-4c34-b773-14364d25cac5/init/0.log" Oct 02 10:15:30 crc kubenswrapper[5035]: I1002 10:15:30.849282 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_046bf1a9-09e7-4ea0-a26e-d6de90838c68/glance-httpd/0.log" Oct 02 10:15:30 crc kubenswrapper[5035]: I1002 10:15:30.867679 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_046bf1a9-09e7-4ea0-a26e-d6de90838c68/glance-log/0.log" Oct 02 10:15:31 crc kubenswrapper[5035]: I1002 10:15:31.025065 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a75ea0f7-0f30-4aa7-8608-78a743ed275b/glance-httpd/0.log" Oct 02 10:15:31 crc kubenswrapper[5035]: I1002 10:15:31.090022 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a75ea0f7-0f30-4aa7-8608-78a743ed275b/glance-log/0.log" Oct 02 10:15:31 crc kubenswrapper[5035]: I1002 10:15:31.261955 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-7b996fb4b6-gvrhc_c596018f-cc4e-4dc4-8c37-def6f2b19f94/init/0.log" Oct 02 10:15:31 crc kubenswrapper[5035]: I1002 10:15:31.594074 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-7b996fb4b6-gvrhc_c596018f-cc4e-4dc4-8c37-def6f2b19f94/ironic-api-log/0.log" Oct 02 10:15:31 crc kubenswrapper[5035]: I1002 10:15:31.599894 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-7b996fb4b6-gvrhc_c596018f-cc4e-4dc4-8c37-def6f2b19f94/init/0.log" Oct 02 10:15:31 crc kubenswrapper[5035]: I1002 10:15:31.659429 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-7b996fb4b6-gvrhc_c596018f-cc4e-4dc4-8c37-def6f2b19f94/ironic-api/0.log" Oct 02 10:15:31 crc kubenswrapper[5035]: I1002 10:15:31.841904 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/init/0.log" Oct 02 10:15:32 crc kubenswrapper[5035]: I1002 10:15:32.015879 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/init/0.log" Oct 02 10:15:32 crc kubenswrapper[5035]: I1002 10:15:32.027593 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/ironic-python-agent-init/0.log" Oct 02 10:15:32 crc kubenswrapper[5035]: I1002 10:15:32.138548 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/ironic-python-agent-init/0.log" Oct 02 10:15:32 crc kubenswrapper[5035]: I1002 10:15:32.352108 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/init/0.log" Oct 02 10:15:32 crc kubenswrapper[5035]: I1002 10:15:32.363924 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/ironic-python-agent-init/0.log" Oct 02 10:15:32 crc kubenswrapper[5035]: I1002 10:15:32.881103 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/init/0.log" Oct 02 10:15:32 crc kubenswrapper[5035]: I1002 10:15:32.881967 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/pxe-init/0.log" Oct 02 10:15:33 crc kubenswrapper[5035]: I1002 10:15:33.075380 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/ironic-python-agent-init/0.log" Oct 02 10:15:33 crc kubenswrapper[5035]: I1002 10:15:33.301939 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/httpboot/0.log" Oct 02 10:15:33 crc kubenswrapper[5035]: I1002 10:15:33.547432 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/ironic-conductor/0.log" Oct 02 10:15:33 crc kubenswrapper[5035]: I1002 10:15:33.641335 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/pxe-init/0.log" Oct 02 10:15:33 crc kubenswrapper[5035]: I1002 10:15:33.751398 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/ramdisk-logs/0.log" Oct 02 10:15:33 crc kubenswrapper[5035]: I1002 10:15:33.798482 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/pxe-init/0.log" Oct 02 10:15:33 crc kubenswrapper[5035]: I1002 10:15:33.981165 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-klp2h_afad3fdc-5375-4bba-9832-f94381ba82aa/init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.070981 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_0307b78d-517d-4442-a0d7-1efa3f46098f/pxe-init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.214597 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-klp2h_afad3fdc-5375-4bba-9832-f94381ba82aa/init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.247437 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-klp2h_afad3fdc-5375-4bba-9832-f94381ba82aa/ironic-db-sync/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.290134 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/ironic-python-agent-init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.432240 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/ironic-python-agent-init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.454147 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/inspector-pxe-init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.495220 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/inspector-pxe-init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.692646 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/inspector-httpboot/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.706131 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/inspector-pxe-init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.728359 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/ironic-python-agent-init/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.762912 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/ironic-inspector/1.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.901502 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/ironic-inspector/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.953971 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/ramdisk-logs/0.log" Oct 02 10:15:34 crc kubenswrapper[5035]: I1002 10:15:34.954773 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_dd4355e8-6dd5-4937-969a-bad47330d25b/ironic-inspector-httpd/0.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.101158 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-db-sync-mmkns_66817970-09ee-4638-a1df-577234b71172/ironic-inspector-db-sync/0.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.149008 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-679cc45cbb-lgkg7_9fec464c-ae40-4d5f-8303-ed2c93babb81/ironic-neutron-agent/2.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.230163 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-679cc45cbb-lgkg7_9fec464c-ae40-4d5f-8303-ed2c93babb81/ironic-neutron-agent/1.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.412865 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-59898c4b58-pz6vt_03747df9-2ce2-42b8-815a-92ffcce73253/keystone-api/0.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.491684 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323321-qsw95_1229f682-2732-4b72-aa97-2443db2ebf64/keystone-cron/0.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.615921 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_e78db9f0-3ff7-444c-9cea-510d690eafc5/kube-state-metrics/0.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.887220 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6f9c95b6cc-5grzm_6cfcc6bb-1b26-42a1-812e-96bd03767545/neutron-api/0.log" Oct 02 10:15:35 crc kubenswrapper[5035]: I1002 10:15:35.933992 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6f9c95b6cc-5grzm_6cfcc6bb-1b26-42a1-812e-96bd03767545/neutron-httpd/0.log" Oct 02 10:15:36 crc kubenswrapper[5035]: I1002 10:15:36.336913 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_75df5c05-833c-4534-8a47-2130520e96ac/nova-api-log/0.log" Oct 02 10:15:36 crc kubenswrapper[5035]: I1002 10:15:36.383624 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_75df5c05-833c-4534-8a47-2130520e96ac/nova-api-api/0.log" Oct 02 10:15:36 crc kubenswrapper[5035]: I1002 10:15:36.675945 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_03343b82-9ffd-4ab8-9b8e-88a1277d3014/nova-cell0-conductor-conductor/0.log" Oct 02 10:15:36 crc kubenswrapper[5035]: I1002 10:15:36.714087 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ad2f225f-00ca-455a-8061-e45abdb56864/nova-cell1-conductor-conductor/0.log" Oct 02 10:15:36 crc kubenswrapper[5035]: I1002 10:15:36.935877 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e316d9ab-3baf-429e-9839-e7dea3125ca4/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 10:15:37 crc kubenswrapper[5035]: I1002 10:15:37.099694 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b803caac-28bd-4f36-b701-e6774f8b83bd/nova-metadata-log/0.log" Oct 02 10:15:37 crc kubenswrapper[5035]: I1002 10:15:37.487836 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_7c61dad9-d722-4d17-958f-594d3d87ca14/nova-scheduler-scheduler/0.log" Oct 02 10:15:37 crc kubenswrapper[5035]: I1002 10:15:37.708337 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9f58a85f-3d79-4ed1-a856-c491969c730f/mysql-bootstrap/0.log" Oct 02 10:15:37 crc kubenswrapper[5035]: I1002 10:15:37.924412 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9f58a85f-3d79-4ed1-a856-c491969c730f/mysql-bootstrap/0.log" Oct 02 10:15:37 crc kubenswrapper[5035]: I1002 10:15:37.951814 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_9f58a85f-3d79-4ed1-a856-c491969c730f/galera/0.log" Oct 02 10:15:38 crc kubenswrapper[5035]: I1002 10:15:38.026088 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b803caac-28bd-4f36-b701-e6774f8b83bd/nova-metadata-metadata/0.log" Oct 02 10:15:38 crc kubenswrapper[5035]: I1002 10:15:38.164696 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_da088dc0-8011-44e7-973b-66e75f33549c/mysql-bootstrap/0.log" Oct 02 10:15:38 crc kubenswrapper[5035]: I1002 10:15:38.517525 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_da088dc0-8011-44e7-973b-66e75f33549c/mysql-bootstrap/0.log" Oct 02 10:15:38 crc kubenswrapper[5035]: I1002 10:15:38.638768 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_da088dc0-8011-44e7-973b-66e75f33549c/galera/0.log" Oct 02 10:15:38 crc kubenswrapper[5035]: I1002 10:15:38.710176 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_68b71f3f-e46f-4025-9935-e000c44652b5/openstackclient/0.log" Oct 02 10:15:38 crc kubenswrapper[5035]: I1002 10:15:38.868309 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jvdts_57ea64fd-e931-45ab-9f0e-678c3c9ba7a2/ovn-controller/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.084433 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-v56nx_92aa4b07-2174-4435-b84c-426260df54fb/openstack-network-exporter/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.201996 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r5m97_89658bb0-5353-4e67-95cd-cdb040086903/ovsdb-server-init/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.475767 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r5m97_89658bb0-5353-4e67-95cd-cdb040086903/ovs-vswitchd/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.477022 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r5m97_89658bb0-5353-4e67-95cd-cdb040086903/ovsdb-server-init/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.521937 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-r5m97_89658bb0-5353-4e67-95cd-cdb040086903/ovsdb-server/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.742618 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3fe804ba-727c-49dd-a12b-43ff832db070/openstack-network-exporter/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.796228 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3fe804ba-727c-49dd-a12b-43ff832db070/ovn-northd/0.log" Oct 02 10:15:39 crc kubenswrapper[5035]: I1002 10:15:39.986728 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_07a982ff-c270-4133-a1a0-612dc7317d4b/openstack-network-exporter/0.log" Oct 02 10:15:40 crc kubenswrapper[5035]: I1002 10:15:40.049686 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_07a982ff-c270-4133-a1a0-612dc7317d4b/ovsdbserver-nb/0.log" Oct 02 10:15:40 crc kubenswrapper[5035]: I1002 10:15:40.210397 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0578ddf1-8297-4720-b0c1-8fa7b7dce11d/openstack-network-exporter/0.log" Oct 02 10:15:40 crc kubenswrapper[5035]: I1002 10:15:40.327731 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_0578ddf1-8297-4720-b0c1-8fa7b7dce11d/ovsdbserver-sb/0.log" Oct 02 10:15:40 crc kubenswrapper[5035]: I1002 10:15:40.512841 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-57bc66b65b-7wmtt_928cad66-0519-4d28-ae2f-a8f5875ac301/placement-api/0.log" Oct 02 10:15:40 crc kubenswrapper[5035]: I1002 10:15:40.631851 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-57bc66b65b-7wmtt_928cad66-0519-4d28-ae2f-a8f5875ac301/placement-log/0.log" Oct 02 10:15:40 crc kubenswrapper[5035]: I1002 10:15:40.812369 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3d7eafce-151f-43f3-8f74-cbfc9f368643/setup-container/0.log" Oct 02 10:15:40 crc kubenswrapper[5035]: I1002 10:15:40.940733 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3d7eafce-151f-43f3-8f74-cbfc9f368643/setup-container/0.log" Oct 02 10:15:41 crc kubenswrapper[5035]: I1002 10:15:41.065125 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3d7eafce-151f-43f3-8f74-cbfc9f368643/rabbitmq/0.log" Oct 02 10:15:41 crc kubenswrapper[5035]: I1002 10:15:41.204917 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d93ca48d-a2a5-4036-930d-948d935cce98/setup-container/0.log" Oct 02 10:15:41 crc kubenswrapper[5035]: I1002 10:15:41.454020 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d93ca48d-a2a5-4036-930d-948d935cce98/setup-container/0.log" Oct 02 10:15:41 crc kubenswrapper[5035]: I1002 10:15:41.491432 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d93ca48d-a2a5-4036-930d-948d935cce98/rabbitmq/0.log" Oct 02 10:15:41 crc kubenswrapper[5035]: I1002 10:15:41.768353 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bc467ccf-p6597_6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36/proxy-httpd/0.log" Oct 02 10:15:41 crc kubenswrapper[5035]: I1002 10:15:41.769225 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6bc467ccf-p6597_6b0fddb8-8e60-4a2b-b4e9-3f7026a55e36/proxy-server/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.011760 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wcbql_65e1909b-5dc0-45ad-af61-f3cd8ce48ace/swift-ring-rebalance/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.050241 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/account-auditor/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.212419 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/account-reaper/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.332226 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/account-replicator/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.349782 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/account-server/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.430134 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/container-auditor/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.581829 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/container-server/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.634713 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/container-replicator/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.682260 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/container-updater/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.826619 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/object-auditor/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.900707 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/object-expirer/0.log" Oct 02 10:15:42 crc kubenswrapper[5035]: I1002 10:15:42.909937 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/object-replicator/0.log" Oct 02 10:15:43 crc kubenswrapper[5035]: I1002 10:15:43.030639 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/object-server/0.log" Oct 02 10:15:43 crc kubenswrapper[5035]: I1002 10:15:43.153681 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/object-updater/0.log" Oct 02 10:15:43 crc kubenswrapper[5035]: I1002 10:15:43.173235 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/rsync/0.log" Oct 02 10:15:43 crc kubenswrapper[5035]: I1002 10:15:43.281311 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_2aaa2a11-2896-4690-819b-61c6069ea6db/swift-recon-cron/0.log" Oct 02 10:15:43 crc kubenswrapper[5035]: I1002 10:15:43.876729 5035 scope.go:117] "RemoveContainer" containerID="c628f3e20bddacc8a7c57934cf125c44f4f12b1d512663a8434c00e6a5a13794" Oct 02 10:15:48 crc kubenswrapper[5035]: I1002 10:15:48.072090 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b08491fc-549c-4176-a8e8-5f733d838e3d/memcached/0.log" Oct 02 10:15:55 crc kubenswrapper[5035]: I1002 10:15:55.538469 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:15:55 crc kubenswrapper[5035]: I1002 10:15:55.539088 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:16:12 crc kubenswrapper[5035]: I1002 10:16:12.213798 5035 generic.go:334] "Generic (PLEG): container finished" podID="1e40b571-0063-4584-b95b-f756e1976f89" containerID="87e5d50e2ac405750b45e0d3f869bfd3e0d7745faf11d8b8de51be31eaa41e4f" exitCode=0 Oct 02 10:16:12 crc kubenswrapper[5035]: I1002 10:16:12.213915 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-xj978" event={"ID":"1e40b571-0063-4584-b95b-f756e1976f89","Type":"ContainerDied","Data":"87e5d50e2ac405750b45e0d3f869bfd3e0d7745faf11d8b8de51be31eaa41e4f"} Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.341282 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.358668 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e40b571-0063-4584-b95b-f756e1976f89-host\") pod \"1e40b571-0063-4584-b95b-f756e1976f89\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.358859 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fkx9\" (UniqueName: \"kubernetes.io/projected/1e40b571-0063-4584-b95b-f756e1976f89-kube-api-access-2fkx9\") pod \"1e40b571-0063-4584-b95b-f756e1976f89\" (UID: \"1e40b571-0063-4584-b95b-f756e1976f89\") " Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.358861 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e40b571-0063-4584-b95b-f756e1976f89-host" (OuterVolumeSpecName: "host") pod "1e40b571-0063-4584-b95b-f756e1976f89" (UID: "1e40b571-0063-4584-b95b-f756e1976f89"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.359619 5035 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e40b571-0063-4584-b95b-f756e1976f89-host\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.368136 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e40b571-0063-4584-b95b-f756e1976f89-kube-api-access-2fkx9" (OuterVolumeSpecName: "kube-api-access-2fkx9") pod "1e40b571-0063-4584-b95b-f756e1976f89" (UID: "1e40b571-0063-4584-b95b-f756e1976f89"). InnerVolumeSpecName "kube-api-access-2fkx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.391222 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q978g/crc-debug-xj978"] Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.403695 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q978g/crc-debug-xj978"] Oct 02 10:16:13 crc kubenswrapper[5035]: I1002 10:16:13.462025 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fkx9\" (UniqueName: \"kubernetes.io/projected/1e40b571-0063-4584-b95b-f756e1976f89-kube-api-access-2fkx9\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.183201 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e40b571-0063-4584-b95b-f756e1976f89" path="/var/lib/kubelet/pods/1e40b571-0063-4584-b95b-f756e1976f89/volumes" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.241271 5035 scope.go:117] "RemoveContainer" containerID="87e5d50e2ac405750b45e0d3f869bfd3e0d7745faf11d8b8de51be31eaa41e4f" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.241361 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-xj978" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.571616 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q978g/crc-debug-x9bf4"] Oct 02 10:16:14 crc kubenswrapper[5035]: E1002 10:16:14.573231 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e40b571-0063-4584-b95b-f756e1976f89" containerName="container-00" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.573363 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e40b571-0063-4584-b95b-f756e1976f89" containerName="container-00" Oct 02 10:16:14 crc kubenswrapper[5035]: E1002 10:16:14.573468 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d060000-b611-4459-b4f4-90dbf4835f12" containerName="collect-profiles" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.573559 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d060000-b611-4459-b4f4-90dbf4835f12" containerName="collect-profiles" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.573857 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d060000-b611-4459-b4f4-90dbf4835f12" containerName="collect-profiles" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.573977 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e40b571-0063-4584-b95b-f756e1976f89" containerName="container-00" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.574758 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.591740 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf95c\" (UniqueName: \"kubernetes.io/projected/b61aa57a-be36-463e-915d-be4529c21509-kube-api-access-lf95c\") pod \"crc-debug-x9bf4\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.592206 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b61aa57a-be36-463e-915d-be4529c21509-host\") pod \"crc-debug-x9bf4\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.693197 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf95c\" (UniqueName: \"kubernetes.io/projected/b61aa57a-be36-463e-915d-be4529c21509-kube-api-access-lf95c\") pod \"crc-debug-x9bf4\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.693422 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b61aa57a-be36-463e-915d-be4529c21509-host\") pod \"crc-debug-x9bf4\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.693562 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b61aa57a-be36-463e-915d-be4529c21509-host\") pod \"crc-debug-x9bf4\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.717927 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf95c\" (UniqueName: \"kubernetes.io/projected/b61aa57a-be36-463e-915d-be4529c21509-kube-api-access-lf95c\") pod \"crc-debug-x9bf4\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:14 crc kubenswrapper[5035]: I1002 10:16:14.891876 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:15 crc kubenswrapper[5035]: I1002 10:16:15.254079 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-x9bf4" event={"ID":"b61aa57a-be36-463e-915d-be4529c21509","Type":"ContainerStarted","Data":"b11ac072e0bb5cc97b64593469f061587ade32689c35ccebe8872f1089665e93"} Oct 02 10:16:15 crc kubenswrapper[5035]: I1002 10:16:15.254739 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-x9bf4" event={"ID":"b61aa57a-be36-463e-915d-be4529c21509","Type":"ContainerStarted","Data":"940520cf9b0a9d9ffb17e3f1766389bddd8dabb8b5f501a5ce0d61600acb0c95"} Oct 02 10:16:15 crc kubenswrapper[5035]: I1002 10:16:15.280949 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-q978g/crc-debug-x9bf4" podStartSLOduration=1.280927737 podStartE2EDuration="1.280927737s" podCreationTimestamp="2025-10-02 10:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 10:16:15.271126011 +0000 UTC m=+2940.627470036" watchObservedRunningTime="2025-10-02 10:16:15.280927737 +0000 UTC m=+2940.637271772" Oct 02 10:16:16 crc kubenswrapper[5035]: I1002 10:16:16.267309 5035 generic.go:334] "Generic (PLEG): container finished" podID="b61aa57a-be36-463e-915d-be4529c21509" containerID="b11ac072e0bb5cc97b64593469f061587ade32689c35ccebe8872f1089665e93" exitCode=0 Oct 02 10:16:16 crc kubenswrapper[5035]: I1002 10:16:16.267413 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-x9bf4" event={"ID":"b61aa57a-be36-463e-915d-be4529c21509","Type":"ContainerDied","Data":"b11ac072e0bb5cc97b64593469f061587ade32689c35ccebe8872f1089665e93"} Oct 02 10:16:17 crc kubenswrapper[5035]: I1002 10:16:17.388072 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:17 crc kubenswrapper[5035]: I1002 10:16:17.437313 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b61aa57a-be36-463e-915d-be4529c21509-host\") pod \"b61aa57a-be36-463e-915d-be4529c21509\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " Oct 02 10:16:17 crc kubenswrapper[5035]: I1002 10:16:17.437410 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b61aa57a-be36-463e-915d-be4529c21509-host" (OuterVolumeSpecName: "host") pod "b61aa57a-be36-463e-915d-be4529c21509" (UID: "b61aa57a-be36-463e-915d-be4529c21509"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:16:17 crc kubenswrapper[5035]: I1002 10:16:17.437602 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf95c\" (UniqueName: \"kubernetes.io/projected/b61aa57a-be36-463e-915d-be4529c21509-kube-api-access-lf95c\") pod \"b61aa57a-be36-463e-915d-be4529c21509\" (UID: \"b61aa57a-be36-463e-915d-be4529c21509\") " Oct 02 10:16:17 crc kubenswrapper[5035]: I1002 10:16:17.437991 5035 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b61aa57a-be36-463e-915d-be4529c21509-host\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:17 crc kubenswrapper[5035]: I1002 10:16:17.450526 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b61aa57a-be36-463e-915d-be4529c21509-kube-api-access-lf95c" (OuterVolumeSpecName: "kube-api-access-lf95c") pod "b61aa57a-be36-463e-915d-be4529c21509" (UID: "b61aa57a-be36-463e-915d-be4529c21509"). InnerVolumeSpecName "kube-api-access-lf95c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:16:17 crc kubenswrapper[5035]: I1002 10:16:17.539276 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf95c\" (UniqueName: \"kubernetes.io/projected/b61aa57a-be36-463e-915d-be4529c21509-kube-api-access-lf95c\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:18 crc kubenswrapper[5035]: I1002 10:16:18.285032 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-x9bf4" event={"ID":"b61aa57a-be36-463e-915d-be4529c21509","Type":"ContainerDied","Data":"940520cf9b0a9d9ffb17e3f1766389bddd8dabb8b5f501a5ce0d61600acb0c95"} Oct 02 10:16:18 crc kubenswrapper[5035]: I1002 10:16:18.285077 5035 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="940520cf9b0a9d9ffb17e3f1766389bddd8dabb8b5f501a5ce0d61600acb0c95" Oct 02 10:16:18 crc kubenswrapper[5035]: I1002 10:16:18.285133 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-x9bf4" Oct 02 10:16:21 crc kubenswrapper[5035]: I1002 10:16:21.031119 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q978g/crc-debug-x9bf4"] Oct 02 10:16:21 crc kubenswrapper[5035]: I1002 10:16:21.041486 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q978g/crc-debug-x9bf4"] Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.176407 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b61aa57a-be36-463e-915d-be4529c21509" path="/var/lib/kubelet/pods/b61aa57a-be36-463e-915d-be4529c21509/volumes" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.240334 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-q978g/crc-debug-gpnj2"] Oct 02 10:16:22 crc kubenswrapper[5035]: E1002 10:16:22.240936 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b61aa57a-be36-463e-915d-be4529c21509" containerName="container-00" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.240975 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b61aa57a-be36-463e-915d-be4529c21509" containerName="container-00" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.241416 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b61aa57a-be36-463e-915d-be4529c21509" containerName="container-00" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.242579 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.317617 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5f91768-640c-4d7e-96e8-71d81807b6fb-host\") pod \"crc-debug-gpnj2\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.317739 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c688v\" (UniqueName: \"kubernetes.io/projected/b5f91768-640c-4d7e-96e8-71d81807b6fb-kube-api-access-c688v\") pod \"crc-debug-gpnj2\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.419200 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c688v\" (UniqueName: \"kubernetes.io/projected/b5f91768-640c-4d7e-96e8-71d81807b6fb-kube-api-access-c688v\") pod \"crc-debug-gpnj2\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.419384 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5f91768-640c-4d7e-96e8-71d81807b6fb-host\") pod \"crc-debug-gpnj2\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.419549 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5f91768-640c-4d7e-96e8-71d81807b6fb-host\") pod \"crc-debug-gpnj2\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.452953 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c688v\" (UniqueName: \"kubernetes.io/projected/b5f91768-640c-4d7e-96e8-71d81807b6fb-kube-api-access-c688v\") pod \"crc-debug-gpnj2\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:22 crc kubenswrapper[5035]: I1002 10:16:22.571291 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:23 crc kubenswrapper[5035]: I1002 10:16:23.334575 5035 generic.go:334] "Generic (PLEG): container finished" podID="b5f91768-640c-4d7e-96e8-71d81807b6fb" containerID="a1fd89dec70b3f7f73bd50b1819b221f5746137dbf2141a4b6cbbd555aed22fb" exitCode=0 Oct 02 10:16:23 crc kubenswrapper[5035]: I1002 10:16:23.334840 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-gpnj2" event={"ID":"b5f91768-640c-4d7e-96e8-71d81807b6fb","Type":"ContainerDied","Data":"a1fd89dec70b3f7f73bd50b1819b221f5746137dbf2141a4b6cbbd555aed22fb"} Oct 02 10:16:23 crc kubenswrapper[5035]: I1002 10:16:23.335151 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/crc-debug-gpnj2" event={"ID":"b5f91768-640c-4d7e-96e8-71d81807b6fb","Type":"ContainerStarted","Data":"b0a81ee04ff38f685e92e44f3ab83b7ae3a544e19acbcf0004248c5da5dd6c75"} Oct 02 10:16:23 crc kubenswrapper[5035]: I1002 10:16:23.384665 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q978g/crc-debug-gpnj2"] Oct 02 10:16:23 crc kubenswrapper[5035]: I1002 10:16:23.393640 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q978g/crc-debug-gpnj2"] Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.482410 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.567741 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5f91768-640c-4d7e-96e8-71d81807b6fb-host\") pod \"b5f91768-640c-4d7e-96e8-71d81807b6fb\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.568107 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c688v\" (UniqueName: \"kubernetes.io/projected/b5f91768-640c-4d7e-96e8-71d81807b6fb-kube-api-access-c688v\") pod \"b5f91768-640c-4d7e-96e8-71d81807b6fb\" (UID: \"b5f91768-640c-4d7e-96e8-71d81807b6fb\") " Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.569794 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5f91768-640c-4d7e-96e8-71d81807b6fb-host" (OuterVolumeSpecName: "host") pod "b5f91768-640c-4d7e-96e8-71d81807b6fb" (UID: "b5f91768-640c-4d7e-96e8-71d81807b6fb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.574876 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5f91768-640c-4d7e-96e8-71d81807b6fb-kube-api-access-c688v" (OuterVolumeSpecName: "kube-api-access-c688v") pod "b5f91768-640c-4d7e-96e8-71d81807b6fb" (UID: "b5f91768-640c-4d7e-96e8-71d81807b6fb"). InnerVolumeSpecName "kube-api-access-c688v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.671034 5035 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5f91768-640c-4d7e-96e8-71d81807b6fb-host\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.671389 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c688v\" (UniqueName: \"kubernetes.io/projected/b5f91768-640c-4d7e-96e8-71d81807b6fb-kube-api-access-c688v\") on node \"crc\" DevicePath \"\"" Oct 02 10:16:24 crc kubenswrapper[5035]: I1002 10:16:24.949950 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf_6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8/util/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.101318 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf_6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8/util/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.124975 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf_6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8/pull/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.141753 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf_6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8/pull/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.296403 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf_6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8/util/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.318655 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf_6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8/pull/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.325963 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7180d9466ff97a79addeaeaaa8a94fe184dffb384df3bdd303962516138bjvf_6b1e8d2e-b8c5-4852-bc3b-09c5a23b36a8/extract/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.359034 5035 scope.go:117] "RemoveContainer" containerID="a1fd89dec70b3f7f73bd50b1819b221f5746137dbf2141a4b6cbbd555aed22fb" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.359090 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/crc-debug-gpnj2" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.508194 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-qmrqc_bb9ed734-53de-48a2-ac4d-57c513e74c83/kube-rbac-proxy/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.521253 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-qmrqc_bb9ed734-53de-48a2-ac4d-57c513e74c83/manager/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.537789 5035 patch_prober.go:28] interesting pod/machine-config-daemon-k6b5v container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.538122 5035 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.538266 5035 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.539261 5035 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3"} pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.539440 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerName="machine-config-daemon" containerID="cri-o://a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" gracePeriod=600 Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.554701 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-msx6h_107fac49-7084-4524-848e-4ebc95387af7/kube-rbac-proxy/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: E1002 10:16:25.668189 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.707441 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-sb4zr_04b44aaf-0053-43c2-a2ad-4cb364e6d045/kube-rbac-proxy/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.762315 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-msx6h_107fac49-7084-4524-848e-4ebc95387af7/manager/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.779761 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-sb4zr_04b44aaf-0053-43c2-a2ad-4cb364e6d045/manager/0.log" Oct 02 10:16:25 crc kubenswrapper[5035]: I1002 10:16:25.866428 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-p72mn_938b87e8-ee53-41d1-b79d-bd57940d12cf/kube-rbac-proxy/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.029240 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-p72mn_938b87e8-ee53-41d1-b79d-bd57940d12cf/manager/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.078624 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-dtzbf_7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5/kube-rbac-proxy/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.125029 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-dtzbf_7ee3db1b-fe71-40b5-8ebb-5cee3301dcd5/manager/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.173021 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5f91768-640c-4d7e-96e8-71d81807b6fb" path="/var/lib/kubelet/pods/b5f91768-640c-4d7e-96e8-71d81807b6fb/volumes" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.219763 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-5znkc_219d8519-1035-4127-a7b3-d59730e87ba8/kube-rbac-proxy/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.264553 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-5znkc_219d8519-1035-4127-a7b3-d59730e87ba8/manager/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.353015 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-hvhzg_2279a6b9-cecc-4b8b-ac52-041a6abe4b9a/kube-rbac-proxy/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.370620 5035 generic.go:334] "Generic (PLEG): container finished" podID="d5f04b76-b225-449c-adc6-e979f26ab0da" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" exitCode=0 Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.370663 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" event={"ID":"d5f04b76-b225-449c-adc6-e979f26ab0da","Type":"ContainerDied","Data":"a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3"} Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.370697 5035 scope.go:117] "RemoveContainer" containerID="668bb58c600ac20dc33ddeae8555bcaa9ee1a27b46e757cd7dc4e3dbed851ad5" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.371341 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:16:26 crc kubenswrapper[5035]: E1002 10:16:26.371725 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.505482 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-d87866488-xc6dw_9c08e6d0-af1a-43cc-a898-d345eba038bb/kube-rbac-proxy/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.521338 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-hvhzg_2279a6b9-cecc-4b8b-ac52-041a6abe4b9a/manager/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.699284 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-hp45m_3bb0cc56-b9e6-42fe-9122-929e1f7a60e8/kube-rbac-proxy/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.700236 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-d87866488-xc6dw_9c08e6d0-af1a-43cc-a898-d345eba038bb/manager/0.log" Oct 02 10:16:26 crc kubenswrapper[5035]: I1002 10:16:26.809107 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-hp45m_3bb0cc56-b9e6-42fe-9122-929e1f7a60e8/manager/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.002822 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-6pm9g_25bfe9fe-5431-49cc-941d-a2bd795bf7d8/kube-rbac-proxy/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.004342 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-6pm9g_25bfe9fe-5431-49cc-941d-a2bd795bf7d8/manager/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.146015 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-bcnlz_98a2c0aa-2954-4b93-be62-0eec45a1e3c4/kube-rbac-proxy/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.237963 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-bcnlz_98a2c0aa-2954-4b93-be62-0eec45a1e3c4/manager/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.295231 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-zh2nk_bf4ed5a3-c0a1-4b98-837b-20de28931ba8/kube-rbac-proxy/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.372439 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-zh2nk_bf4ed5a3-c0a1-4b98-837b-20de28931ba8/manager/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.427012 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-sgg6c_1dc0efdb-db90-4467-9255-056281840a4a/kube-rbac-proxy/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.559429 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-nsrkb_ac688684-8289-4fe0-a776-14ec42952e8f/kube-rbac-proxy/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.600963 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-sgg6c_1dc0efdb-db90-4467-9255-056281840a4a/manager/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.619351 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-nsrkb_ac688684-8289-4fe0-a776-14ec42952e8f/manager/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.729146 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-9grgw_db1e3e37-5811-4be9-bbcb-62ae86c895fd/kube-rbac-proxy/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.771467 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-9grgw_db1e3e37-5811-4be9-bbcb-62ae86c895fd/manager/0.log" Oct 02 10:16:27 crc kubenswrapper[5035]: I1002 10:16:27.917127 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65bbb5d7c7-shdfq_0a35394d-10b4-408d-8666-780370f58653/kube-rbac-proxy/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.022383 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5578564d9c-6vmnb_db3c6e4e-7e62-40dc-91fa-0c79a1eb8010/kube-rbac-proxy/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.220404 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-sz275_8c739198-67cc-4ffe-92ee-4ec30714256f/registry-server/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.311257 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5578564d9c-6vmnb_db3c6e4e-7e62-40dc-91fa-0c79a1eb8010/operator/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.480424 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-rs2lh_b9fa1e3a-3fc9-46e6-a1fb-e51755336945/kube-rbac-proxy/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.499028 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-rs2lh_b9fa1e3a-3fc9-46e6-a1fb-e51755336945/manager/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.643633 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-8f4pt_0b78dcac-905a-43ee-bfaa-ec62206f67ba/kube-rbac-proxy/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.659039 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65bbb5d7c7-shdfq_0a35394d-10b4-408d-8666-780370f58653/manager/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.692869 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-8f4pt_0b78dcac-905a-43ee-bfaa-ec62206f67ba/manager/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.832617 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-2lqs2_4f7f3b68-a6ec-42b1-be68-d027bc3b5012/kube-rbac-proxy/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.834919 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-nj2jw_3b85995e-f7f9-42d1-9cd5-c32d2ba5444a/operator/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.876672 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-2lqs2_4f7f3b68-a6ec-42b1-be68-d027bc3b5012/manager/0.log" Oct 02 10:16:28 crc kubenswrapper[5035]: I1002 10:16:28.979626 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-49mnw_1b7f13af-0af5-4ffd-b677-e583068cd598/kube-rbac-proxy/0.log" Oct 02 10:16:29 crc kubenswrapper[5035]: I1002 10:16:29.040889 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-49mnw_1b7f13af-0af5-4ffd-b677-e583068cd598/manager/0.log" Oct 02 10:16:29 crc kubenswrapper[5035]: I1002 10:16:29.066145 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-msfdb_c7a8160d-c732-4736-a1b0-57999db04976/kube-rbac-proxy/0.log" Oct 02 10:16:29 crc kubenswrapper[5035]: I1002 10:16:29.117634 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-msfdb_c7a8160d-c732-4736-a1b0-57999db04976/manager/0.log" Oct 02 10:16:29 crc kubenswrapper[5035]: I1002 10:16:29.196358 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-m58vc_099bc72a-55f2-438f-b90c-5e54cc4e7f8e/kube-rbac-proxy/0.log" Oct 02 10:16:29 crc kubenswrapper[5035]: I1002 10:16:29.217382 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-m58vc_099bc72a-55f2-438f-b90c-5e54cc4e7f8e/manager/0.log" Oct 02 10:16:39 crc kubenswrapper[5035]: I1002 10:16:39.163510 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:16:39 crc kubenswrapper[5035]: E1002 10:16:39.164240 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:16:44 crc kubenswrapper[5035]: I1002 10:16:44.468278 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-k2d9m_4a7a7830-3a3d-4571-9dcd-ec0280e87e6a/control-plane-machine-set-operator/0.log" Oct 02 10:16:44 crc kubenswrapper[5035]: I1002 10:16:44.626687 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9qdrb_368eaba6-c405-4ba6-afa9-13b32093851a/kube-rbac-proxy/0.log" Oct 02 10:16:44 crc kubenswrapper[5035]: I1002 10:16:44.639469 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9qdrb_368eaba6-c405-4ba6-afa9-13b32093851a/machine-api-operator/0.log" Oct 02 10:16:50 crc kubenswrapper[5035]: I1002 10:16:50.164267 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:16:50 crc kubenswrapper[5035]: E1002 10:16:50.165241 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:16:56 crc kubenswrapper[5035]: I1002 10:16:56.685454 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-plbbh_ab058dbd-cd9f-4754-a021-d16a6351fac2/cert-manager-controller/0.log" Oct 02 10:16:56 crc kubenswrapper[5035]: I1002 10:16:56.798497 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-86j6v_d2cd272f-9d31-4957-80b7-977c88b98ad0/cert-manager-cainjector/0.log" Oct 02 10:16:56 crc kubenswrapper[5035]: I1002 10:16:56.844602 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-p2tjr_3af23dbf-8290-4edc-b019-e17fcffb8489/cert-manager-webhook/0.log" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.095583 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-62czh"] Oct 02 10:16:57 crc kubenswrapper[5035]: E1002 10:16:57.096354 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f91768-640c-4d7e-96e8-71d81807b6fb" containerName="container-00" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.096465 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f91768-640c-4d7e-96e8-71d81807b6fb" containerName="container-00" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.096859 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5f91768-640c-4d7e-96e8-71d81807b6fb" containerName="container-00" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.100364 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.109441 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62czh"] Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.206032 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk5xk\" (UniqueName: \"kubernetes.io/projected/3a0322f8-b51b-4d01-aa37-bcb861b8401f-kube-api-access-jk5xk\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.206154 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-utilities\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.206249 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-catalog-content\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.308643 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-utilities\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.309056 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-catalog-content\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.309324 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk5xk\" (UniqueName: \"kubernetes.io/projected/3a0322f8-b51b-4d01-aa37-bcb861b8401f-kube-api-access-jk5xk\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.310235 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-utilities\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.310292 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-catalog-content\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.334335 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk5xk\" (UniqueName: \"kubernetes.io/projected/3a0322f8-b51b-4d01-aa37-bcb861b8401f-kube-api-access-jk5xk\") pod \"certified-operators-62czh\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.427629 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:16:57 crc kubenswrapper[5035]: I1002 10:16:57.918616 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62czh"] Oct 02 10:16:58 crc kubenswrapper[5035]: I1002 10:16:58.678287 5035 generic.go:334] "Generic (PLEG): container finished" podID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerID="5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34" exitCode=0 Oct 02 10:16:58 crc kubenswrapper[5035]: I1002 10:16:58.678382 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62czh" event={"ID":"3a0322f8-b51b-4d01-aa37-bcb861b8401f","Type":"ContainerDied","Data":"5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34"} Oct 02 10:16:58 crc kubenswrapper[5035]: I1002 10:16:58.678663 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62czh" event={"ID":"3a0322f8-b51b-4d01-aa37-bcb861b8401f","Type":"ContainerStarted","Data":"26fba50b397832f311500d84077b1584b1d23406c940d99ec0a5b957ad1e141c"} Oct 02 10:17:00 crc kubenswrapper[5035]: I1002 10:17:00.696111 5035 generic.go:334] "Generic (PLEG): container finished" podID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerID="d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989" exitCode=0 Oct 02 10:17:00 crc kubenswrapper[5035]: I1002 10:17:00.696240 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62czh" event={"ID":"3a0322f8-b51b-4d01-aa37-bcb861b8401f","Type":"ContainerDied","Data":"d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989"} Oct 02 10:17:01 crc kubenswrapper[5035]: I1002 10:17:01.163217 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:17:01 crc kubenswrapper[5035]: E1002 10:17:01.163846 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:17:01 crc kubenswrapper[5035]: I1002 10:17:01.721260 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62czh" event={"ID":"3a0322f8-b51b-4d01-aa37-bcb861b8401f","Type":"ContainerStarted","Data":"1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33"} Oct 02 10:17:01 crc kubenswrapper[5035]: I1002 10:17:01.746164 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-62czh" podStartSLOduration=2.302329769 podStartE2EDuration="4.746146732s" podCreationTimestamp="2025-10-02 10:16:57 +0000 UTC" firstStartedPulling="2025-10-02 10:16:58.680927188 +0000 UTC m=+2984.037271213" lastFinishedPulling="2025-10-02 10:17:01.124744151 +0000 UTC m=+2986.481088176" observedRunningTime="2025-10-02 10:17:01.74198129 +0000 UTC m=+2987.098325305" watchObservedRunningTime="2025-10-02 10:17:01.746146732 +0000 UTC m=+2987.102490757" Oct 02 10:17:07 crc kubenswrapper[5035]: I1002 10:17:07.429127 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:17:07 crc kubenswrapper[5035]: I1002 10:17:07.429652 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:17:07 crc kubenswrapper[5035]: I1002 10:17:07.515587 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:17:07 crc kubenswrapper[5035]: I1002 10:17:07.840440 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:17:07 crc kubenswrapper[5035]: I1002 10:17:07.891622 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62czh"] Oct 02 10:17:09 crc kubenswrapper[5035]: I1002 10:17:09.274139 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-5l6pn_5dc578e1-5285-4af1-a1f4-23f30e092e1b/nmstate-console-plugin/0.log" Oct 02 10:17:09 crc kubenswrapper[5035]: I1002 10:17:09.425841 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-mk96w_aa6d0e0d-c7f8-4967-9142-4ef34426f311/nmstate-handler/0.log" Oct 02 10:17:09 crc kubenswrapper[5035]: I1002 10:17:09.496190 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-pdzhf_193c7ee9-3048-41dd-8987-d6ad9ff934fc/nmstate-metrics/0.log" Oct 02 10:17:09 crc kubenswrapper[5035]: I1002 10:17:09.522013 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-pdzhf_193c7ee9-3048-41dd-8987-d6ad9ff934fc/kube-rbac-proxy/0.log" Oct 02 10:17:09 crc kubenswrapper[5035]: I1002 10:17:09.673463 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-brkcf_de005a53-14f0-49a4-9fbf-492c4b466dcd/nmstate-operator/0.log" Oct 02 10:17:09 crc kubenswrapper[5035]: I1002 10:17:09.728816 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-hwqtt_b0ccfcb3-f1bb-4573-8106-c730f6b7920c/nmstate-webhook/0.log" Oct 02 10:17:09 crc kubenswrapper[5035]: I1002 10:17:09.800323 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-62czh" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="registry-server" containerID="cri-o://1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33" gracePeriod=2 Oct 02 10:17:09 crc kubenswrapper[5035]: E1002 10:17:09.857689 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a0322f8_b51b_4d01_aa37_bcb861b8401f.slice/crio-conmon-1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.264732 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.342983 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-catalog-content\") pod \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.343161 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk5xk\" (UniqueName: \"kubernetes.io/projected/3a0322f8-b51b-4d01-aa37-bcb861b8401f-kube-api-access-jk5xk\") pod \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.343234 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-utilities\") pod \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\" (UID: \"3a0322f8-b51b-4d01-aa37-bcb861b8401f\") " Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.344673 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-utilities" (OuterVolumeSpecName: "utilities") pod "3a0322f8-b51b-4d01-aa37-bcb861b8401f" (UID: "3a0322f8-b51b-4d01-aa37-bcb861b8401f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.348961 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a0322f8-b51b-4d01-aa37-bcb861b8401f-kube-api-access-jk5xk" (OuterVolumeSpecName: "kube-api-access-jk5xk") pod "3a0322f8-b51b-4d01-aa37-bcb861b8401f" (UID: "3a0322f8-b51b-4d01-aa37-bcb861b8401f"). InnerVolumeSpecName "kube-api-access-jk5xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.424489 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a0322f8-b51b-4d01-aa37-bcb861b8401f" (UID: "3a0322f8-b51b-4d01-aa37-bcb861b8401f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.444755 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.444789 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk5xk\" (UniqueName: \"kubernetes.io/projected/3a0322f8-b51b-4d01-aa37-bcb861b8401f-kube-api-access-jk5xk\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.444803 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a0322f8-b51b-4d01-aa37-bcb861b8401f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.809163 5035 generic.go:334] "Generic (PLEG): container finished" podID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerID="1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33" exitCode=0 Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.809205 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62czh" event={"ID":"3a0322f8-b51b-4d01-aa37-bcb861b8401f","Type":"ContainerDied","Data":"1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33"} Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.809231 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62czh" event={"ID":"3a0322f8-b51b-4d01-aa37-bcb861b8401f","Type":"ContainerDied","Data":"26fba50b397832f311500d84077b1584b1d23406c940d99ec0a5b957ad1e141c"} Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.809230 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62czh" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.809248 5035 scope.go:117] "RemoveContainer" containerID="1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.837014 5035 scope.go:117] "RemoveContainer" containerID="d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.853250 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62czh"] Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.862480 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-62czh"] Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.899619 5035 scope.go:117] "RemoveContainer" containerID="5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.920548 5035 scope.go:117] "RemoveContainer" containerID="1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33" Oct 02 10:17:10 crc kubenswrapper[5035]: E1002 10:17:10.921008 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33\": container with ID starting with 1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33 not found: ID does not exist" containerID="1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.921045 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33"} err="failed to get container status \"1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33\": rpc error: code = NotFound desc = could not find container \"1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33\": container with ID starting with 1b82a3994bba78e48ced33015c094df548554d1cc466617b0bfaec4a394d8a33 not found: ID does not exist" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.921069 5035 scope.go:117] "RemoveContainer" containerID="d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989" Oct 02 10:17:10 crc kubenswrapper[5035]: E1002 10:17:10.921438 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989\": container with ID starting with d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989 not found: ID does not exist" containerID="d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.921493 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989"} err="failed to get container status \"d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989\": rpc error: code = NotFound desc = could not find container \"d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989\": container with ID starting with d99786aecb4d867d8eeaef7d12f48bba0bce7cc0b32f4f5e2daa9fc232498989 not found: ID does not exist" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.921523 5035 scope.go:117] "RemoveContainer" containerID="5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34" Oct 02 10:17:10 crc kubenswrapper[5035]: E1002 10:17:10.921983 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34\": container with ID starting with 5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34 not found: ID does not exist" containerID="5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34" Oct 02 10:17:10 crc kubenswrapper[5035]: I1002 10:17:10.922014 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34"} err="failed to get container status \"5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34\": rpc error: code = NotFound desc = could not find container \"5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34\": container with ID starting with 5214bc843c19a45c496e45b8693cdb077a714d318f6f4139a65f64c53937bc34 not found: ID does not exist" Oct 02 10:17:12 crc kubenswrapper[5035]: I1002 10:17:12.173942 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" path="/var/lib/kubelet/pods/3a0322f8-b51b-4d01-aa37-bcb861b8401f/volumes" Oct 02 10:17:14 crc kubenswrapper[5035]: I1002 10:17:14.163206 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:17:14 crc kubenswrapper[5035]: E1002 10:17:14.163685 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.149762 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-sms6d_ebcd114a-13c7-4c49-a893-7016d73a6e0c/kube-rbac-proxy/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.323086 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-sms6d_ebcd114a-13c7-4c49-a893-7016d73a6e0c/controller/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.372796 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-frr-files/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.540571 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-frr-files/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.585918 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-reloader/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.625799 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-reloader/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.629587 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-metrics/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.927510 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-metrics/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.933447 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-frr-files/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.968069 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-metrics/0.log" Oct 02 10:17:23 crc kubenswrapper[5035]: I1002 10:17:23.977223 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-reloader/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.134875 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-reloader/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.138427 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-metrics/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.154497 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/cp-frr-files/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.182004 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/controller/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.335234 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/frr-metrics/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.356572 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/kube-rbac-proxy/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.394687 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/kube-rbac-proxy-frr/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.569876 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-9qrxg_3b2dae08-474c-4f53-802a-35133f5f5119/frr-k8s-webhook-server/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.614479 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/reloader/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.797026 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-b9b6fcb4f-9qjh9_e83c1515-25dc-45f1-9ff6-547382cce9b3/manager/0.log" Oct 02 10:17:24 crc kubenswrapper[5035]: I1002 10:17:24.986562 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7cb8bcccbf-94b58_cdbb31cd-2a07-479e-aebe-55a862352dc9/webhook-server/0.log" Oct 02 10:17:25 crc kubenswrapper[5035]: I1002 10:17:25.058521 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-kn2pw_aa3805ad-fdc2-4c45-9c0e-6c0d933d8351/kube-rbac-proxy/0.log" Oct 02 10:17:25 crc kubenswrapper[5035]: I1002 10:17:25.255689 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-576rd_4536e046-502a-4344-87d4-d05e53f3a73b/frr/0.log" Oct 02 10:17:25 crc kubenswrapper[5035]: I1002 10:17:25.452195 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-kn2pw_aa3805ad-fdc2-4c45-9c0e-6c0d933d8351/speaker/0.log" Oct 02 10:17:27 crc kubenswrapper[5035]: I1002 10:17:27.164033 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:17:27 crc kubenswrapper[5035]: E1002 10:17:27.165095 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.391998 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48_c643d696-0eaa-4495-acfe-6035dd231a0c/util/0.log" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.522490 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48_c643d696-0eaa-4495-acfe-6035dd231a0c/util/0.log" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.557066 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48_c643d696-0eaa-4495-acfe-6035dd231a0c/pull/0.log" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.587383 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48_c643d696-0eaa-4495-acfe-6035dd231a0c/pull/0.log" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.726044 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48_c643d696-0eaa-4495-acfe-6035dd231a0c/util/0.log" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.758610 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48_c643d696-0eaa-4495-acfe-6035dd231a0c/pull/0.log" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.764493 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2zcm48_c643d696-0eaa-4495-acfe-6035dd231a0c/extract/0.log" Oct 02 10:17:38 crc kubenswrapper[5035]: I1002 10:17:38.876619 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9gx62_31404512-afbe-4f3e-838a-2d46075439ef/extract-utilities/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.015183 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9gx62_31404512-afbe-4f3e-838a-2d46075439ef/extract-utilities/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.050886 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9gx62_31404512-afbe-4f3e-838a-2d46075439ef/extract-content/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.081951 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9gx62_31404512-afbe-4f3e-838a-2d46075439ef/extract-content/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.209132 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9gx62_31404512-afbe-4f3e-838a-2d46075439ef/extract-content/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.220845 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9gx62_31404512-afbe-4f3e-838a-2d46075439ef/extract-utilities/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.457432 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qdmjm_2939c1ba-1ace-4aea-b0ff-ca389a25c3e1/extract-utilities/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.594332 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-9gx62_31404512-afbe-4f3e-838a-2d46075439ef/registry-server/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.654971 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qdmjm_2939c1ba-1ace-4aea-b0ff-ca389a25c3e1/extract-utilities/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.669745 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qdmjm_2939c1ba-1ace-4aea-b0ff-ca389a25c3e1/extract-content/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.684752 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qdmjm_2939c1ba-1ace-4aea-b0ff-ca389a25c3e1/extract-content/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.799225 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qdmjm_2939c1ba-1ace-4aea-b0ff-ca389a25c3e1/extract-utilities/0.log" Oct 02 10:17:39 crc kubenswrapper[5035]: I1002 10:17:39.843957 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qdmjm_2939c1ba-1ace-4aea-b0ff-ca389a25c3e1/extract-content/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.034148 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq_d61d4b64-e37d-4da5-9b5b-3ec51b94e790/util/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.164583 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:17:40 crc kubenswrapper[5035]: E1002 10:17:40.164889 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.266551 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-qdmjm_2939c1ba-1ace-4aea-b0ff-ca389a25c3e1/registry-server/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.270229 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq_d61d4b64-e37d-4da5-9b5b-3ec51b94e790/pull/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.273212 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq_d61d4b64-e37d-4da5-9b5b-3ec51b94e790/pull/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.300853 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq_d61d4b64-e37d-4da5-9b5b-3ec51b94e790/util/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.490670 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq_d61d4b64-e37d-4da5-9b5b-3ec51b94e790/pull/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.506746 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq_d61d4b64-e37d-4da5-9b5b-3ec51b94e790/util/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.552297 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c9qkpq_d61d4b64-e37d-4da5-9b5b-3ec51b94e790/extract/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.733290 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mslb8_859b8e87-5dc4-47fb-8c0b-17bce445deae/extract-utilities/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.753231 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-jrspr_95bd2b11-6306-4c82-be67-be8d03769473/marketplace-operator/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.886981 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mslb8_859b8e87-5dc4-47fb-8c0b-17bce445deae/extract-utilities/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.916117 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mslb8_859b8e87-5dc4-47fb-8c0b-17bce445deae/extract-content/0.log" Oct 02 10:17:40 crc kubenswrapper[5035]: I1002 10:17:40.928873 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mslb8_859b8e87-5dc4-47fb-8c0b-17bce445deae/extract-content/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.076117 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mslb8_859b8e87-5dc4-47fb-8c0b-17bce445deae/extract-utilities/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.078626 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mslb8_859b8e87-5dc4-47fb-8c0b-17bce445deae/extract-content/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.186990 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mslb8_859b8e87-5dc4-47fb-8c0b-17bce445deae/registry-server/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.260733 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p2xvl_b65c1e0e-ed37-4471-b369-6aa6c863288e/extract-utilities/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.419873 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p2xvl_b65c1e0e-ed37-4471-b369-6aa6c863288e/extract-content/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.421087 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p2xvl_b65c1e0e-ed37-4471-b369-6aa6c863288e/extract-content/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.424037 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p2xvl_b65c1e0e-ed37-4471-b369-6aa6c863288e/extract-utilities/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.603259 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p2xvl_b65c1e0e-ed37-4471-b369-6aa6c863288e/extract-utilities/0.log" Oct 02 10:17:41 crc kubenswrapper[5035]: I1002 10:17:41.619082 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p2xvl_b65c1e0e-ed37-4471-b369-6aa6c863288e/extract-content/0.log" Oct 02 10:17:42 crc kubenswrapper[5035]: I1002 10:17:42.097137 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p2xvl_b65c1e0e-ed37-4471-b369-6aa6c863288e/registry-server/0.log" Oct 02 10:17:54 crc kubenswrapper[5035]: I1002 10:17:54.163234 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:17:54 crc kubenswrapper[5035]: E1002 10:17:54.164071 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:18:08 crc kubenswrapper[5035]: I1002 10:18:08.162835 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:18:08 crc kubenswrapper[5035]: E1002 10:18:08.163796 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:18:14 crc kubenswrapper[5035]: E1002 10:18:14.733109 5035 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.154:40764->38.102.83.154:40175: write tcp 38.102.83.154:40764->38.102.83.154:40175: write: broken pipe Oct 02 10:18:22 crc kubenswrapper[5035]: I1002 10:18:22.163613 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:18:22 crc kubenswrapper[5035]: E1002 10:18:22.165126 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:18:34 crc kubenswrapper[5035]: I1002 10:18:34.164499 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:18:34 crc kubenswrapper[5035]: E1002 10:18:34.167521 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:18:47 crc kubenswrapper[5035]: I1002 10:18:47.163458 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:18:47 crc kubenswrapper[5035]: E1002 10:18:47.164201 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:18:58 crc kubenswrapper[5035]: I1002 10:18:58.171725 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:18:58 crc kubenswrapper[5035]: E1002 10:18:58.172851 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.541931 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2zvfm"] Oct 02 10:19:03 crc kubenswrapper[5035]: E1002 10:19:03.543268 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="extract-content" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.543287 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="extract-content" Oct 02 10:19:03 crc kubenswrapper[5035]: E1002 10:19:03.543320 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="extract-utilities" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.543332 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="extract-utilities" Oct 02 10:19:03 crc kubenswrapper[5035]: E1002 10:19:03.543349 5035 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="registry-server" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.543359 5035 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="registry-server" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.543634 5035 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a0322f8-b51b-4d01-aa37-bcb861b8401f" containerName="registry-server" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.546643 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.572781 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2zvfm"] Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.645369 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-utilities\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.645520 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h9qq\" (UniqueName: \"kubernetes.io/projected/a8f27855-5565-4dbe-9128-0d69b8c70c01-kube-api-access-7h9qq\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.645575 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-catalog-content\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.738481 5035 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xswdq"] Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.760780 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.767053 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xswdq"] Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.791855 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-utilities\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.792022 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h9qq\" (UniqueName: \"kubernetes.io/projected/a8f27855-5565-4dbe-9128-0d69b8c70c01-kube-api-access-7h9qq\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.792059 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-catalog-content\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.792764 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-catalog-content\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.792997 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-utilities\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.814996 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h9qq\" (UniqueName: \"kubernetes.io/projected/a8f27855-5565-4dbe-9128-0d69b8c70c01-kube-api-access-7h9qq\") pod \"community-operators-2zvfm\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.880608 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.894106 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-utilities\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.894206 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-catalog-content\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.894300 5035 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8cj9\" (UniqueName: \"kubernetes.io/projected/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-kube-api-access-q8cj9\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.996338 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8cj9\" (UniqueName: \"kubernetes.io/projected/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-kube-api-access-q8cj9\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.996708 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-utilities\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.996765 5035 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-catalog-content\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.997308 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-utilities\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:03 crc kubenswrapper[5035]: I1002 10:19:03.997382 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-catalog-content\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.015050 5035 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8cj9\" (UniqueName: \"kubernetes.io/projected/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-kube-api-access-q8cj9\") pod \"redhat-operators-xswdq\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.080025 5035 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.525502 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2zvfm"] Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.634260 5035 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xswdq"] Oct 02 10:19:04 crc kubenswrapper[5035]: W1002 10:19:04.650314 5035 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48d48ef8_53ec_4260_b8be_8d0bfed82d8b.slice/crio-48e95565609946284a963f5f0f1862bd00b9b8c224b12fcbca1efd59ff779d64 WatchSource:0}: Error finding container 48e95565609946284a963f5f0f1862bd00b9b8c224b12fcbca1efd59ff779d64: Status 404 returned error can't find the container with id 48e95565609946284a963f5f0f1862bd00b9b8c224b12fcbca1efd59ff779d64 Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.974509 5035 generic.go:334] "Generic (PLEG): container finished" podID="a8f27855-5565-4dbe-9128-0d69b8c70c01" containerID="a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd" exitCode=0 Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.974585 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zvfm" event={"ID":"a8f27855-5565-4dbe-9128-0d69b8c70c01","Type":"ContainerDied","Data":"a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd"} Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.974613 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zvfm" event={"ID":"a8f27855-5565-4dbe-9128-0d69b8c70c01","Type":"ContainerStarted","Data":"41ffe431e5b3dcd92e2d08d42e399cd458b54d63475037c77661c1da8a8bc33b"} Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.976339 5035 generic.go:334] "Generic (PLEG): container finished" podID="48d48ef8-53ec-4260-b8be-8d0bfed82d8b" containerID="5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099" exitCode=0 Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.976364 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xswdq" event={"ID":"48d48ef8-53ec-4260-b8be-8d0bfed82d8b","Type":"ContainerDied","Data":"5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099"} Oct 02 10:19:04 crc kubenswrapper[5035]: I1002 10:19:04.976379 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xswdq" event={"ID":"48d48ef8-53ec-4260-b8be-8d0bfed82d8b","Type":"ContainerStarted","Data":"48e95565609946284a963f5f0f1862bd00b9b8c224b12fcbca1efd59ff779d64"} Oct 02 10:19:06 crc kubenswrapper[5035]: I1002 10:19:06.013105 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xswdq" event={"ID":"48d48ef8-53ec-4260-b8be-8d0bfed82d8b","Type":"ContainerStarted","Data":"93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58"} Oct 02 10:19:06 crc kubenswrapper[5035]: I1002 10:19:06.033148 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zvfm" event={"ID":"a8f27855-5565-4dbe-9128-0d69b8c70c01","Type":"ContainerStarted","Data":"12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3"} Oct 02 10:19:07 crc kubenswrapper[5035]: I1002 10:19:07.050912 5035 generic.go:334] "Generic (PLEG): container finished" podID="48d48ef8-53ec-4260-b8be-8d0bfed82d8b" containerID="93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58" exitCode=0 Oct 02 10:19:07 crc kubenswrapper[5035]: I1002 10:19:07.051352 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xswdq" event={"ID":"48d48ef8-53ec-4260-b8be-8d0bfed82d8b","Type":"ContainerDied","Data":"93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58"} Oct 02 10:19:07 crc kubenswrapper[5035]: I1002 10:19:07.054794 5035 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 10:19:07 crc kubenswrapper[5035]: I1002 10:19:07.061168 5035 generic.go:334] "Generic (PLEG): container finished" podID="a8f27855-5565-4dbe-9128-0d69b8c70c01" containerID="12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3" exitCode=0 Oct 02 10:19:07 crc kubenswrapper[5035]: I1002 10:19:07.061223 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zvfm" event={"ID":"a8f27855-5565-4dbe-9128-0d69b8c70c01","Type":"ContainerDied","Data":"12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3"} Oct 02 10:19:08 crc kubenswrapper[5035]: I1002 10:19:08.072940 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xswdq" event={"ID":"48d48ef8-53ec-4260-b8be-8d0bfed82d8b","Type":"ContainerStarted","Data":"a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956"} Oct 02 10:19:08 crc kubenswrapper[5035]: I1002 10:19:08.079968 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zvfm" event={"ID":"a8f27855-5565-4dbe-9128-0d69b8c70c01","Type":"ContainerStarted","Data":"3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be"} Oct 02 10:19:08 crc kubenswrapper[5035]: I1002 10:19:08.107111 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xswdq" podStartSLOduration=2.40958979 podStartE2EDuration="5.107081905s" podCreationTimestamp="2025-10-02 10:19:03 +0000 UTC" firstStartedPulling="2025-10-02 10:19:04.977560875 +0000 UTC m=+3110.333904900" lastFinishedPulling="2025-10-02 10:19:07.67505298 +0000 UTC m=+3113.031397015" observedRunningTime="2025-10-02 10:19:08.098607928 +0000 UTC m=+3113.454951963" watchObservedRunningTime="2025-10-02 10:19:08.107081905 +0000 UTC m=+3113.463425930" Oct 02 10:19:08 crc kubenswrapper[5035]: I1002 10:19:08.125150 5035 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2zvfm" podStartSLOduration=2.53360863 podStartE2EDuration="5.125129012s" podCreationTimestamp="2025-10-02 10:19:03 +0000 UTC" firstStartedPulling="2025-10-02 10:19:04.975866666 +0000 UTC m=+3110.332210691" lastFinishedPulling="2025-10-02 10:19:07.567387038 +0000 UTC m=+3112.923731073" observedRunningTime="2025-10-02 10:19:08.122124434 +0000 UTC m=+3113.478468459" watchObservedRunningTime="2025-10-02 10:19:08.125129012 +0000 UTC m=+3113.481473037" Oct 02 10:19:10 crc kubenswrapper[5035]: I1002 10:19:10.163964 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:19:10 crc kubenswrapper[5035]: E1002 10:19:10.164589 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:19:13 crc kubenswrapper[5035]: I1002 10:19:13.880880 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:13 crc kubenswrapper[5035]: I1002 10:19:13.881474 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:13 crc kubenswrapper[5035]: I1002 10:19:13.949169 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:14 crc kubenswrapper[5035]: I1002 10:19:14.081112 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:14 crc kubenswrapper[5035]: I1002 10:19:14.081174 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:14 crc kubenswrapper[5035]: I1002 10:19:14.158426 5035 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:14 crc kubenswrapper[5035]: I1002 10:19:14.220910 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:14 crc kubenswrapper[5035]: I1002 10:19:14.236214 5035 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.003028 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2zvfm"] Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.174844 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2zvfm" podUID="a8f27855-5565-4dbe-9128-0d69b8c70c01" containerName="registry-server" containerID="cri-o://3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be" gracePeriod=2 Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.606361 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xswdq"] Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.607285 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xswdq" podUID="48d48ef8-53ec-4260-b8be-8d0bfed82d8b" containerName="registry-server" containerID="cri-o://a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956" gracePeriod=2 Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.826370 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.963632 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-catalog-content\") pod \"a8f27855-5565-4dbe-9128-0d69b8c70c01\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.963697 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h9qq\" (UniqueName: \"kubernetes.io/projected/a8f27855-5565-4dbe-9128-0d69b8c70c01-kube-api-access-7h9qq\") pod \"a8f27855-5565-4dbe-9128-0d69b8c70c01\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.963735 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-utilities\") pod \"a8f27855-5565-4dbe-9128-0d69b8c70c01\" (UID: \"a8f27855-5565-4dbe-9128-0d69b8c70c01\") " Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.964911 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-utilities" (OuterVolumeSpecName: "utilities") pod "a8f27855-5565-4dbe-9128-0d69b8c70c01" (UID: "a8f27855-5565-4dbe-9128-0d69b8c70c01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:19:16 crc kubenswrapper[5035]: I1002 10:19:16.985761 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f27855-5565-4dbe-9128-0d69b8c70c01-kube-api-access-7h9qq" (OuterVolumeSpecName: "kube-api-access-7h9qq") pod "a8f27855-5565-4dbe-9128-0d69b8c70c01" (UID: "a8f27855-5565-4dbe-9128-0d69b8c70c01"). InnerVolumeSpecName "kube-api-access-7h9qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.066823 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h9qq\" (UniqueName: \"kubernetes.io/projected/a8f27855-5565-4dbe-9128-0d69b8c70c01-kube-api-access-7h9qq\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.067742 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.078543 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.169192 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-catalog-content\") pod \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.169233 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8cj9\" (UniqueName: \"kubernetes.io/projected/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-kube-api-access-q8cj9\") pod \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.169328 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-utilities\") pod \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\" (UID: \"48d48ef8-53ec-4260-b8be-8d0bfed82d8b\") " Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.170849 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-utilities" (OuterVolumeSpecName: "utilities") pod "48d48ef8-53ec-4260-b8be-8d0bfed82d8b" (UID: "48d48ef8-53ec-4260-b8be-8d0bfed82d8b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.186696 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-kube-api-access-q8cj9" (OuterVolumeSpecName: "kube-api-access-q8cj9") pod "48d48ef8-53ec-4260-b8be-8d0bfed82d8b" (UID: "48d48ef8-53ec-4260-b8be-8d0bfed82d8b"). InnerVolumeSpecName "kube-api-access-q8cj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.197797 5035 generic.go:334] "Generic (PLEG): container finished" podID="48d48ef8-53ec-4260-b8be-8d0bfed82d8b" containerID="a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956" exitCode=0 Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.197889 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xswdq" event={"ID":"48d48ef8-53ec-4260-b8be-8d0bfed82d8b","Type":"ContainerDied","Data":"a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956"} Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.197928 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xswdq" event={"ID":"48d48ef8-53ec-4260-b8be-8d0bfed82d8b","Type":"ContainerDied","Data":"48e95565609946284a963f5f0f1862bd00b9b8c224b12fcbca1efd59ff779d64"} Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.197956 5035 scope.go:117] "RemoveContainer" containerID="a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.198115 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xswdq" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.203441 5035 generic.go:334] "Generic (PLEG): container finished" podID="a8f27855-5565-4dbe-9128-0d69b8c70c01" containerID="3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be" exitCode=0 Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.203488 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zvfm" event={"ID":"a8f27855-5565-4dbe-9128-0d69b8c70c01","Type":"ContainerDied","Data":"3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be"} Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.203518 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2zvfm" event={"ID":"a8f27855-5565-4dbe-9128-0d69b8c70c01","Type":"ContainerDied","Data":"41ffe431e5b3dcd92e2d08d42e399cd458b54d63475037c77661c1da8a8bc33b"} Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.203622 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2zvfm" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.221761 5035 scope.go:117] "RemoveContainer" containerID="93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.239330 5035 scope.go:117] "RemoveContainer" containerID="5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.250055 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48d48ef8-53ec-4260-b8be-8d0bfed82d8b" (UID: "48d48ef8-53ec-4260-b8be-8d0bfed82d8b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.259254 5035 scope.go:117] "RemoveContainer" containerID="a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956" Oct 02 10:19:17 crc kubenswrapper[5035]: E1002 10:19:17.260192 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956\": container with ID starting with a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956 not found: ID does not exist" containerID="a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.260231 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956"} err="failed to get container status \"a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956\": rpc error: code = NotFound desc = could not find container \"a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956\": container with ID starting with a2331e3b69a033879aed6d520e25c2a703091676bb960553b20758ccb3458956 not found: ID does not exist" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.260250 5035 scope.go:117] "RemoveContainer" containerID="93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58" Oct 02 10:19:17 crc kubenswrapper[5035]: E1002 10:19:17.260580 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58\": container with ID starting with 93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58 not found: ID does not exist" containerID="93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.260620 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58"} err="failed to get container status \"93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58\": rpc error: code = NotFound desc = could not find container \"93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58\": container with ID starting with 93388287a77e19fa98f36abe937c4b0b626b91b8676dba588c95298952c5be58 not found: ID does not exist" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.260664 5035 scope.go:117] "RemoveContainer" containerID="5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099" Oct 02 10:19:17 crc kubenswrapper[5035]: E1002 10:19:17.261008 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099\": container with ID starting with 5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099 not found: ID does not exist" containerID="5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.261041 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099"} err="failed to get container status \"5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099\": rpc error: code = NotFound desc = could not find container \"5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099\": container with ID starting with 5324c068506cc89255768b55d1c6a50caa57e81fdc57ddd6f6096b2654b21099 not found: ID does not exist" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.261061 5035 scope.go:117] "RemoveContainer" containerID="3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.272795 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.272825 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8cj9\" (UniqueName: \"kubernetes.io/projected/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-kube-api-access-q8cj9\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.272841 5035 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d48ef8-53ec-4260-b8be-8d0bfed82d8b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.291493 5035 scope.go:117] "RemoveContainer" containerID="12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.390796 5035 scope.go:117] "RemoveContainer" containerID="a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.473983 5035 scope.go:117] "RemoveContainer" containerID="3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be" Oct 02 10:19:17 crc kubenswrapper[5035]: E1002 10:19:17.474709 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be\": container with ID starting with 3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be not found: ID does not exist" containerID="3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.474746 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be"} err="failed to get container status \"3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be\": rpc error: code = NotFound desc = could not find container \"3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be\": container with ID starting with 3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be not found: ID does not exist" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.474771 5035 scope.go:117] "RemoveContainer" containerID="12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3" Oct 02 10:19:17 crc kubenswrapper[5035]: E1002 10:19:17.476759 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3\": container with ID starting with 12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3 not found: ID does not exist" containerID="12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.476803 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3"} err="failed to get container status \"12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3\": rpc error: code = NotFound desc = could not find container \"12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3\": container with ID starting with 12f72b256f3d94a9974d3a90ee44f3455a60e2bc9e02bd26b2058a119cab12a3 not found: ID does not exist" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.476827 5035 scope.go:117] "RemoveContainer" containerID="a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd" Oct 02 10:19:17 crc kubenswrapper[5035]: E1002 10:19:17.477260 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd\": container with ID starting with a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd not found: ID does not exist" containerID="a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.477284 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd"} err="failed to get container status \"a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd\": rpc error: code = NotFound desc = could not find container \"a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd\": container with ID starting with a4adccebd8c5e7b5acd260ec195e78355d549914ba3066cf8b58ae44b48f2ccd not found: ID does not exist" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.542030 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xswdq"] Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.548268 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xswdq"] Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.609312 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8f27855-5565-4dbe-9128-0d69b8c70c01" (UID: "a8f27855-5565-4dbe-9128-0d69b8c70c01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.680261 5035 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8f27855-5565-4dbe-9128-0d69b8c70c01-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.846150 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2zvfm"] Oct 02 10:19:17 crc kubenswrapper[5035]: I1002 10:19:17.855476 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2zvfm"] Oct 02 10:19:18 crc kubenswrapper[5035]: I1002 10:19:18.186806 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d48ef8-53ec-4260-b8be-8d0bfed82d8b" path="/var/lib/kubelet/pods/48d48ef8-53ec-4260-b8be-8d0bfed82d8b/volumes" Oct 02 10:19:18 crc kubenswrapper[5035]: I1002 10:19:18.188342 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8f27855-5565-4dbe-9128-0d69b8c70c01" path="/var/lib/kubelet/pods/a8f27855-5565-4dbe-9128-0d69b8c70c01/volumes" Oct 02 10:19:23 crc kubenswrapper[5035]: I1002 10:19:23.164038 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:19:23 crc kubenswrapper[5035]: E1002 10:19:23.165199 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:19:23 crc kubenswrapper[5035]: E1002 10:19:23.514710 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8f27855_5565_4dbe_9128_0d69b8c70c01.slice/crio-3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:19:24 crc kubenswrapper[5035]: I1002 10:19:24.299126 5035 generic.go:334] "Generic (PLEG): container finished" podID="183893c4-7c7f-4e8a-80a0-5850149d5f9c" containerID="89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b" exitCode=0 Oct 02 10:19:24 crc kubenswrapper[5035]: I1002 10:19:24.299217 5035 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-q978g/must-gather-8ncgt" event={"ID":"183893c4-7c7f-4e8a-80a0-5850149d5f9c","Type":"ContainerDied","Data":"89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b"} Oct 02 10:19:24 crc kubenswrapper[5035]: I1002 10:19:24.300119 5035 scope.go:117] "RemoveContainer" containerID="89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b" Oct 02 10:19:24 crc kubenswrapper[5035]: I1002 10:19:24.461425 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q978g_must-gather-8ncgt_183893c4-7c7f-4e8a-80a0-5850149d5f9c/gather/0.log" Oct 02 10:19:32 crc kubenswrapper[5035]: I1002 10:19:32.659636 5035 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-q978g/must-gather-8ncgt"] Oct 02 10:19:32 crc kubenswrapper[5035]: I1002 10:19:32.660474 5035 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-q978g/must-gather-8ncgt" podUID="183893c4-7c7f-4e8a-80a0-5850149d5f9c" containerName="copy" containerID="cri-o://0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2" gracePeriod=2 Oct 02 10:19:32 crc kubenswrapper[5035]: I1002 10:19:32.673600 5035 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-q978g/must-gather-8ncgt"] Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.102263 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q978g_must-gather-8ncgt_183893c4-7c7f-4e8a-80a0-5850149d5f9c/copy/0.log" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.103237 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.240183 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ng6s6\" (UniqueName: \"kubernetes.io/projected/183893c4-7c7f-4e8a-80a0-5850149d5f9c-kube-api-access-ng6s6\") pod \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.240345 5035 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/183893c4-7c7f-4e8a-80a0-5850149d5f9c-must-gather-output\") pod \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\" (UID: \"183893c4-7c7f-4e8a-80a0-5850149d5f9c\") " Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.255858 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/183893c4-7c7f-4e8a-80a0-5850149d5f9c-kube-api-access-ng6s6" (OuterVolumeSpecName: "kube-api-access-ng6s6") pod "183893c4-7c7f-4e8a-80a0-5850149d5f9c" (UID: "183893c4-7c7f-4e8a-80a0-5850149d5f9c"). InnerVolumeSpecName "kube-api-access-ng6s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.342723 5035 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ng6s6\" (UniqueName: \"kubernetes.io/projected/183893c4-7c7f-4e8a-80a0-5850149d5f9c-kube-api-access-ng6s6\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.375943 5035 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/183893c4-7c7f-4e8a-80a0-5850149d5f9c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "183893c4-7c7f-4e8a-80a0-5850149d5f9c" (UID: "183893c4-7c7f-4e8a-80a0-5850149d5f9c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.399899 5035 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-q978g_must-gather-8ncgt_183893c4-7c7f-4e8a-80a0-5850149d5f9c/copy/0.log" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.400171 5035 generic.go:334] "Generic (PLEG): container finished" podID="183893c4-7c7f-4e8a-80a0-5850149d5f9c" containerID="0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2" exitCode=143 Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.400231 5035 scope.go:117] "RemoveContainer" containerID="0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.400232 5035 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-q978g/must-gather-8ncgt" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.418443 5035 scope.go:117] "RemoveContainer" containerID="89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.444980 5035 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/183893c4-7c7f-4e8a-80a0-5850149d5f9c-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.478214 5035 scope.go:117] "RemoveContainer" containerID="0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2" Oct 02 10:19:33 crc kubenswrapper[5035]: E1002 10:19:33.478726 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2\": container with ID starting with 0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2 not found: ID does not exist" containerID="0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.478768 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2"} err="failed to get container status \"0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2\": rpc error: code = NotFound desc = could not find container \"0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2\": container with ID starting with 0f2eda9b6588e72ac55142b3716fce086e51172e10c4c819fcfa5d1d1d77d6e2 not found: ID does not exist" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.478795 5035 scope.go:117] "RemoveContainer" containerID="89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b" Oct 02 10:19:33 crc kubenswrapper[5035]: E1002 10:19:33.479225 5035 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b\": container with ID starting with 89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b not found: ID does not exist" containerID="89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b" Oct 02 10:19:33 crc kubenswrapper[5035]: I1002 10:19:33.479269 5035 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b"} err="failed to get container status \"89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b\": rpc error: code = NotFound desc = could not find container \"89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b\": container with ID starting with 89a9349ec83e5a7f997e9021c31efd0d84b8389ea2f9d5214d590e01578a2e5b not found: ID does not exist" Oct 02 10:19:33 crc kubenswrapper[5035]: E1002 10:19:33.780315 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8f27855_5565_4dbe_9128_0d69b8c70c01.slice/crio-3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:19:34 crc kubenswrapper[5035]: I1002 10:19:34.180089 5035 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="183893c4-7c7f-4e8a-80a0-5850149d5f9c" path="/var/lib/kubelet/pods/183893c4-7c7f-4e8a-80a0-5850149d5f9c/volumes" Oct 02 10:19:36 crc kubenswrapper[5035]: I1002 10:19:36.175751 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:19:36 crc kubenswrapper[5035]: E1002 10:19:36.176396 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:19:44 crc kubenswrapper[5035]: E1002 10:19:44.058035 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8f27855_5565_4dbe_9128_0d69b8c70c01.slice/crio-3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:19:49 crc kubenswrapper[5035]: I1002 10:19:49.162826 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:19:49 crc kubenswrapper[5035]: E1002 10:19:49.163460 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:19:54 crc kubenswrapper[5035]: E1002 10:19:54.383559 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8f27855_5565_4dbe_9128_0d69b8c70c01.slice/crio-3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:20:03 crc kubenswrapper[5035]: I1002 10:20:03.163737 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:20:03 crc kubenswrapper[5035]: E1002 10:20:03.164622 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:20:04 crc kubenswrapper[5035]: E1002 10:20:04.635116 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8f27855_5565_4dbe_9128_0d69b8c70c01.slice/crio-3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:20:14 crc kubenswrapper[5035]: E1002 10:20:14.910208 5035 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8f27855_5565_4dbe_9128_0d69b8c70c01.slice/crio-3adfbc15b7d327625e8861c4281e08e6518a2891892d546b530bf497e02e30be.scope\": RecentStats: unable to find data in memory cache]" Oct 02 10:20:15 crc kubenswrapper[5035]: I1002 10:20:15.163988 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:20:15 crc kubenswrapper[5035]: E1002 10:20:15.164834 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:20:29 crc kubenswrapper[5035]: I1002 10:20:29.163521 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:20:29 crc kubenswrapper[5035]: E1002 10:20:29.164349 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:20:40 crc kubenswrapper[5035]: I1002 10:20:40.163874 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:20:40 crc kubenswrapper[5035]: E1002 10:20:40.164845 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" Oct 02 10:20:52 crc kubenswrapper[5035]: I1002 10:20:52.163660 5035 scope.go:117] "RemoveContainer" containerID="a82aba68af214e0ade4a9a05d1f9d63e925e90ebf9a9945f23f156dab408eab3" Oct 02 10:20:52 crc kubenswrapper[5035]: E1002 10:20:52.165267 5035 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-k6b5v_openshift-machine-config-operator(d5f04b76-b225-449c-adc6-e979f26ab0da)\"" pod="openshift-machine-config-operator/machine-config-daemon-k6b5v" podUID="d5f04b76-b225-449c-adc6-e979f26ab0da" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067451020024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067451021017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067442220016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067442221015460 5ustar corecore